[27/51] [partial] hbase-site git commit: Published site at 620d70d6186fb800299bcc62ad7179fccfd1be41.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html b/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html index 9913946..055a934 100644 --- a/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html +++ b/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html @@ -236,18 +236,22 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -org.apache.hadoop.hbase.tool +org.apache.hadoop.hbase.thrift2.client -org.apache.hadoop.hbase.util +org.apache.hadoop.hbase.tool -org.apache.hadoop.hbase.wal +org.apache.hadoop.hbase.util +org.apache.hadoop.hbase.wal + + + org.apache.hadoop.hbase.zookeeper @@ -1948,6 +1952,40 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. + + + +Uses of Pair in org.apache.hadoop.hbase.thrift2.client + +Methods in org.apache.hadoop.hbase.thrift2.client that return Pair + +Modifier and Type +Method and Description + + + +Pairhttps://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer,https://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer +ThriftAdmin.getAlterStatus(byte[]tableName) + + +Pairhttps://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer,https://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer +ThriftAdmin.getAlterStatus(TableNametableName) + + +Pairorg.apache.hadoop.hbase.thrift2.generated.THBaseService.Client,org.apache.thrift.transport.TTransport +ThriftConnection.DefaultThriftClientBuilder.getClient() + + +Pairorg.apache.hadoop.hbase.thrift2.generated.THBaseService.Client,org.apache.thrift.transport.TTransport +ThriftConnection.HTTPThriftClientBuilder.getClient() + + +abstract Pairorg.apache.hadoop.hbase.thrift2.generated.THBaseService.Client,org.apache.thrift.transport.TTransport +ThriftClientBuilder.getClient() + + + + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/org/apache/hadoop/hbase/util/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html index 25ec8fe..4e1d4c5 100644 --- a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html @@ -541,14 +541,14 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.util.PrettyPrinter.Unit +org.apache.hadoop.hbase.util.IdReadWriteLock.ReferenceType +org.apache.hadoop.hbase.util.ChecksumType org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE +org.apache.hadoop.hbase.util.PrettyPrinter.Unit +org.apache.hadoop.hbase.util.PoolMap.PoolType +org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.PureJavaComparer (implements org.apache.hadoop.hbase.util.Bytes.ComparerT) org.apache.hadoop.hbase.util.Order org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.UnsafeComparer (implements org.apache.hadoop.hbase.util.Bytes.ComparerT) -org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.PureJavaComparer (implements org.apache.hadoop.hbase.util.Bytes.ComparerT) -org.apache.hadoop.hbase.util.PoolMap.PoolType -org.apache.hadoop.hbase.util.ChecksumType -org.apache.hadoop.hbase.util.IdReadWriteLock.ReferenceType http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/org/apache/hadoop/hbase/util/package-use.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/util/package-use.html b/devapidocs/org/apache/hadoop/hbase/util/package-use.html index 6336a97..7c797bf 100644 --- a/devapidocs/org/apache/hadoop/hbase/util/package-use.html +++ b/devapidocs/org/apache/hadoop/hbase/util/package-use.html @@ -311,14 +311,18 @@ service. -org.apache.hadoop.hbase.tool +org.apache.hadoop.hbase.thrift2.client -org.apache.hadoop.hbase.tool.coprocessor +org.apache.hadoop.hbase.tool +org.apache.hadoop.hbase.tool.coprocessor + + + org.apache.hadoop.hbase.types @@ -326,19 +330,19 @@ service. extensible data type API.
[27/51] [partial] hbase-site git commit: Published site at 281d6429e55149cc4c05430dcc1d1dc136d8b245.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/901d593a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html b/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html index ff0e4d5..c3c0cb1 100644 --- a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html +++ b/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html @@ -761,58 +761,65 @@ default void +MasterObserver.postIsRpcThrottleEnabled(ObserverContextMasterCoprocessorEnvironmentctx, +booleanrpcThrottleEnabled) +Called after getting if is rpc throttle enabled. + + + +default void MasterObserver.postListDecommissionedRegionServers(ObserverContextMasterCoprocessorEnvironmentctx) Called after list decommissioned region servers. - + default void MasterObserver.postListNamespaceDescriptors(ObserverContextMasterCoprocessorEnvironmentctx, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListNamespaceDescriptordescriptors) Called after a listNamespaceDescriptors request has been processed. - + default void MasterObserver.postListReplicationPeers(ObserverContextMasterCoprocessorEnvironmentctx, https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringregex) Called after list replication peers. - + default void MasterObserver.postListRSGroups(ObserverContextMasterCoprocessorEnvironmentctx) Called after listing region server group information. - + default void MasterObserver.postListSnapshot(ObserverContextMasterCoprocessorEnvironmentctx, SnapshotDescriptionsnapshot) Called after listSnapshots request has been processed. - + default void MasterObserver.postLockHeartbeat(ObserverContextMasterCoprocessorEnvironmentctx) Called after heartbeat to a lock. - + default void RegionObserver.postMemStoreCompaction(ObserverContextRegionCoprocessorEnvironmentc, Storestore) Called after the in memory compaction is finished. - + default void MasterObserver.postMergeRegions(ObserverContextMasterCoprocessorEnvironmentc, RegionInfo[]regionsToMerge) called after merge regions request. - + default void MasterObserver.postMergeRegionsCommitAction(ObserverContextMasterCoprocessorEnvironmentctx, RegionInfo[]regionsToMerge, @@ -820,7 +827,7 @@ This will be called after META step as part of regions merge transaction. - + default void MasterObserver.postModifyNamespace(ObserverContextMasterCoprocessorEnvironmentctx, NamespaceDescriptorcurrentNsDescriptor) @@ -829,7 +836,7 @@ - + default void MasterObserver.postModifyNamespace(ObserverContextMasterCoprocessorEnvironmentctx, NamespaceDescriptoroldNsDescriptor, @@ -837,7 +844,7 @@ Called after the modifyNamespace operation has been requested. - + default void MasterObserver.postModifyTable(ObserverContextMasterCoprocessorEnvironmentctx, TableNametableName, @@ -847,7 +854,7 @@ - + default void MasterObserver.postModifyTable(ObserverContextMasterCoprocessorEnvironmentctx, TableNametableName, @@ -856,7 +863,7 @@ Called after the modifyTable operation has been requested. - + default void MasterObserver.postMove(ObserverContextMasterCoprocessorEnvironmentctx, RegionInforegion, @@ -865,7 +872,7 @@ Called after the region move has been requested. - + default void MasterObserver.postMoveServers(ObserverContextMasterCoprocessorEnvironmentctx, https://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">SetAddressservers, @@ -873,7 +880,7 @@ Called after servers are moved to target region server group - + default void MasterObserver.postMoveServersAndTables(ObserverContextMasterCoprocessorEnvironmentctx, https://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">SetAddressservers, @@ -882,7 +889,7 @@ Called after servers are moved to target region server group - + default void MasterObserver.postMoveTables(ObserverContextMasterCoprocessorEnvironmentctx, https://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">SetTableNametables, @@ -890,7 +897,7 @@ Called after servers are moved to target region server group - + default Cell RegionObserver.postMutationBeforeWAL(ObserverContextRegionCoprocessorEnvironmentctx,
[27/51] [partial] hbase-site git commit: Published site at 466fa920fee572fe20db3b77ebf539dc304d5f31.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2bf59208/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncBatchRpcRetryingCaller.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncBatchRpcRetryingCaller.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncBatchRpcRetryingCaller.html index cd0ff28..4f9947f 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncBatchRpcRetryingCaller.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncBatchRpcRetryingCaller.html @@ -31,161 +31,161 @@ 023import static org.apache.hadoop.hbase.client.ConnectionUtils.resetController; 024import static org.apache.hadoop.hbase.client.ConnectionUtils.translateException; 025import static org.apache.hadoop.hbase.util.CollectionUtils.computeIfAbsent; -026 -027import org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer; -028 -029import java.io.IOException; -030import java.util.ArrayList; -031import java.util.Collections; -032import java.util.HashMap; -033import java.util.IdentityHashMap; -034import java.util.List; -035import java.util.Map; -036import java.util.Optional; -037import java.util.concurrent.CompletableFuture; -038import java.util.concurrent.ConcurrentHashMap; -039import java.util.concurrent.ConcurrentLinkedQueue; -040import java.util.concurrent.ConcurrentMap; -041import java.util.concurrent.ConcurrentSkipListMap; -042import java.util.concurrent.TimeUnit; -043import java.util.function.Supplier; -044import java.util.stream.Collectors; -045import java.util.stream.Stream; -046 -047import org.apache.hadoop.hbase.CellScannable; -048import org.apache.hadoop.hbase.DoNotRetryIOException; -049import org.apache.hadoop.hbase.HRegionLocation; -050import org.apache.hadoop.hbase.ServerName; -051import org.apache.hadoop.hbase.TableName; -052import org.apache.yetus.audience.InterfaceAudience; -053import org.slf4j.Logger; -054import org.slf4j.LoggerFactory; -055import org.apache.hadoop.hbase.client.MultiResponse.RegionResult; -056import org.apache.hadoop.hbase.client.RetriesExhaustedException.ThrowableWithExtraContext; -057import org.apache.hadoop.hbase.ipc.HBaseRpcController; -058import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter; -059import org.apache.hadoop.hbase.shaded.protobuf.ResponseConverter; -060import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos; -061import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ClientService; -062import org.apache.hadoop.hbase.util.Bytes; -063import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -064 -065/** -066 * Retry caller for batch. -067 * p -068 * Notice that, the {@link #operationTimeoutNs} is the total time limit now which is the same with -069 * other single operations -070 * p -071 * And the {@link #maxAttempts} is a limit for each single operation in the batch logically. In the -072 * implementation, we will record a {@code tries} parameter for each operation group, and if it is -073 * split to several groups when retrying, the sub groups will inherit the {@code tries}. You can -074 * imagine that the whole retrying process is a tree, and the {@link #maxAttempts} is the limit of -075 * the depth of the tree. -076 */ -077@InterfaceAudience.Private -078class AsyncBatchRpcRetryingCallerT { -079 -080 private static final Logger LOG = LoggerFactory.getLogger(AsyncBatchRpcRetryingCaller.class); -081 -082 private final HashedWheelTimer retryTimer; -083 -084 private final AsyncConnectionImpl conn; -085 -086 private final TableName tableName; -087 -088 private final ListAction actions; -089 -090 private final ListCompletableFutureT futures; -091 -092 private final IdentityHashMapAction, CompletableFutureT action2Future; -093 -094 private final IdentityHashMapAction, ListThrowableWithExtraContext action2Errors; -095 -096 private final long pauseNs; -097 -098 private final int maxAttempts; -099 -100 private final long operationTimeoutNs; -101 -102 private final long rpcTimeoutNs; -103 -104 private final int startLogErrorsCnt; -105 -106 private final long startNs; -107 -108 // we can not use HRegionLocation as the map key because the hashCode and equals method of -109 // HRegionLocation only consider serverName. -110 private static final class RegionRequest { -111 -112public final HRegionLocation loc; -113 -114public final ConcurrentLinkedQueueAction actions = new ConcurrentLinkedQueue(); -115 -116public RegionRequest(HRegionLocation loc) { -117 this.loc = loc; -118} -119 } -120 -121 private static final class ServerRequest { -122 -123public final ConcurrentMapbyte[], RegionRequest actionsByRegion = -124new ConcurrentSkipListMap(Bytes.BYTES_COMPARATOR); -125 -126public void addAction(HRegionLocation loc, Action action) { -127 computeIfAbsent(actionsByRegion,
[27/51] [partial] hbase-site git commit: Published site at e4b6b4afb933a961f543537875f87a2dc62d3757.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperation.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperation.html b/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperation.html index c4e8c8b..aa58108 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperation.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.ObserverOperation.html @@ -82,602 +82,613 @@ 074 public static final String USER_COPROCESSORS_ENABLED_CONF_KEY = 075"hbase.coprocessor.user.enabled"; 076 public static final boolean DEFAULT_USER_COPROCESSORS_ENABLED = true; -077 -078 private static final Logger LOG = LoggerFactory.getLogger(CoprocessorHost.class); -079 protected Abortable abortable; -080 /** Ordered set of loaded coprocessors with lock */ -081 protected final SortedListE coprocEnvironments = -082 new SortedList(new EnvironmentPriorityComparator()); -083 protected Configuration conf; -084 // unique file prefix to use for local copies of jars when classloading -085 protected String pathPrefix; -086 protected AtomicInteger loadSequence = new AtomicInteger(); -087 -088 public CoprocessorHost(Abortable abortable) { -089this.abortable = abortable; -090this.pathPrefix = UUID.randomUUID().toString(); -091 } -092 -093 /** -094 * Not to be confused with the per-object _coprocessors_ (above), -095 * coprocessorNames is static and stores the set of all coprocessors ever -096 * loaded by any thread in this JVM. It is strictly additive: coprocessors are -097 * added to coprocessorNames, by checkAndLoadInstance() but are never removed, since -098 * the intention is to preserve a history of all loaded coprocessors for -099 * diagnosis in case of server crash (HBASE-4014). -100 */ -101 private static SetString coprocessorNames = -102 Collections.synchronizedSet(new HashSetString()); -103 -104 public static SetString getLoadedCoprocessors() { -105synchronized (coprocessorNames) { -106 return new HashSet(coprocessorNames); -107} -108 } -109 -110 /** -111 * Used to create a parameter to the HServerLoad constructor so that -112 * HServerLoad can provide information about the coprocessors loaded by this -113 * regionserver. -114 * (HBASE-4070: Improve region server metrics to report loaded coprocessors -115 * to master). -116 */ -117 public SetString getCoprocessors() { -118SetString returnValue = new TreeSet(); -119for (E e: coprocEnvironments) { -120 returnValue.add(e.getInstance().getClass().getSimpleName()); -121} -122return returnValue; -123 } -124 -125 /** -126 * Load system coprocessors once only. Read the class names from configuration. -127 * Called by constructor. -128 */ -129 protected void loadSystemCoprocessors(Configuration conf, String confKey) { -130boolean coprocessorsEnabled = conf.getBoolean(COPROCESSORS_ENABLED_CONF_KEY, -131 DEFAULT_COPROCESSORS_ENABLED); -132if (!coprocessorsEnabled) { -133 return; -134} -135 -136Class? implClass; -137 -138// load default coprocessors from configure file -139String[] defaultCPClasses = conf.getStrings(confKey); -140if (defaultCPClasses == null || defaultCPClasses.length == 0) -141 return; -142 -143int priority = Coprocessor.PRIORITY_SYSTEM; -144for (String className : defaultCPClasses) { -145 className = className.trim(); -146 if (findCoprocessor(className) != null) { -147// If already loaded will just continue -148LOG.warn("Attempted duplicate loading of " + className + "; skipped"); -149continue; -150 } -151 ClassLoader cl = this.getClass().getClassLoader(); -152 Thread.currentThread().setContextClassLoader(cl); -153 try { -154implClass = cl.loadClass(className); -155// Add coprocessors as we go to guard against case where a coprocessor is specified twice -156// in the configuration -157E env = checkAndLoadInstance(implClass, priority, conf); -158if (env != null) { -159 this.coprocEnvironments.add(env); -160 LOG.info("System coprocessor {} loaded, priority={}.", className, priority); -161 ++priority; -162} -163 } catch (Throwable t) { -164// We always abort if system coprocessors cannot be loaded -165abortServer(className, t); -166 } -167} -168 } -169 -170 /** -171 * Load a coprocessor implementation into the host -172 * @param path path to implementation jar -173 * @param className the main class name -174 * @param priority chaining priority -175 * @param conf configuration for coprocessor
[27/51] [partial] hbase-site git commit: Published site at 3ab895979b643a2980bcdb7fee2078f14b614210.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionFactory.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionFactory.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionFactory.html index ee11a41..79ad648 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionFactory.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionFactory.html @@ -120,6 +120,6 @@ -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceState.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceState.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceState.html index 85b5b1b..ae5fa8d 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceState.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceState.html @@ -191,6 +191,6 @@ -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceStubMaker.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceStubMaker.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceStubMaker.html index 581a1bf..306c08d 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceStubMaker.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.MasterServiceStubMaker.html @@ -120,6 +120,6 @@ -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.ServerErrors.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.ServerErrors.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.ServerErrors.html index dd01782..cebfb9d 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.ServerErrors.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.ServerErrors.html @@ -162,6 +162,6 @@ -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.html index c6dda63..f1d19b2 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.ServerErrorTracker.html @@ -183,6 +183,6 @@ -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/ConnectionImplementation.html
[27/51] [partial] hbase-site git commit: Published site at 7820ba1dbdba58b1002cdfde08eb21aa7a0bb6da.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/BloomType.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/BloomType.html b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/BloomType.html index 90e5a60..56edaab 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/BloomType.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/BloomType.html @@ -115,6 +115,13 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. +org.apache.hadoop.hbase.thrift2 + +Provides an HBase http://thrift.apache.org/;>Thrift +service. + + + org.apache.hadoop.hbase.util @@ -428,6 +435,37 @@ the order they are declared. + + + +Uses of BloomType in org.apache.hadoop.hbase.thrift2 + +Methods in org.apache.hadoop.hbase.thrift2 that return BloomType + +Modifier and Type +Method and Description + + + +static BloomType +ThriftUtilities.bloomFilterFromThrift(org.apache.hadoop.hbase.thrift2.generated.TBloomFilterTypein) + + + + +Methods in org.apache.hadoop.hbase.thrift2 with parameters of type BloomType + +Modifier and Type +Method and Description + + + +static org.apache.hadoop.hbase.thrift2.generated.TBloomFilterType +ThriftUtilities.bloomFilterFromHBase(BloomTypein) + + + + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html index 3e892ef..9384f68 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegion.html @@ -405,7 +405,7 @@ HRegionServer.getRegionByEncodedName(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringencodedRegionName) -(package private) static HRegion +static HRegion HRegion.newHRegion(org.apache.hadoop.fs.PathtableDir, WALwal, org.apache.hadoop.fs.FileSystemfs, http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/RegionServerServices.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/RegionServerServices.html b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/RegionServerServices.html index b06ddd9..bc92a6b 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/RegionServerServices.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/RegionServerServices.html @@ -606,7 +606,7 @@ -(package private) static HRegion +static HRegion HRegion.newHRegion(org.apache.hadoop.fs.PathtableDir, WALwal, org.apache.hadoop.fs.FileSystemfs, http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html index 8b58771..50c6f2d 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html @@ -716,20 +716,20 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status -org.apache.hadoop.hbase.regionserver.ScanType -org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection -org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType -org.apache.hadoop.hbase.regionserver.Region.Operation -org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType org.apache.hadoop.hbase.regionserver.ScannerContext.NextState -org.apache.hadoop.hbase.regionserver.BloomType +org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status +org.apache.hadoop.hbase.regionserver.ChunkCreator.ChunkType +org.apache.hadoop.hbase.regionserver.TimeRangeTracker.Type org.apache.hadoop.hbase.regionserver.HRegion.FlushResult.Result
[27/51] [partial] hbase-site git commit: Published site at c448604ceb987d113913f0583452b2abce04db0d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html index b8d0e70..ac5997f 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -private static class HRegionServer.CompactionChecker +private static class HRegionServer.CompactionChecker extends ScheduledChore @@ -233,7 +233,7 @@ extends instance -private finalHRegionServer instance +private finalHRegionServer instance @@ -242,7 +242,7 @@ extends majorCompactPriority -private finalint majorCompactPriority +private finalint majorCompactPriority @@ -251,7 +251,7 @@ extends DEFAULT_PRIORITY -private static finalint DEFAULT_PRIORITY +private static finalint DEFAULT_PRIORITY See Also: Constant Field Values @@ -264,7 +264,7 @@ extends iteration -privatelong iteration +privatelong iteration @@ -281,7 +281,7 @@ extends CompactionChecker -CompactionChecker(HRegionServerh, +CompactionChecker(HRegionServerh, intsleepTime, Stoppablestopper) @@ -300,7 +300,7 @@ extends chore -protectedvoidchore() +protectedvoidchore() Description copied from class:ScheduledChore The task to execute on each scheduled execution of the Chore http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html index 5d4def3..30395a4 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html @@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab"; -private static class HRegionServer.MovedRegionInfo +private static class HRegionServer.MovedRegionInfo extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object @@ -218,7 +218,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html serverName -private finalServerName serverName +private finalServerName serverName @@ -227,7 +227,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html seqNum -private finallong seqNum +private finallong seqNum @@ -236,7 +236,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html ts -private finallong ts +private finallong ts @@ -253,7 +253,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html MovedRegionInfo -publicMovedRegionInfo(ServerNameserverName, +publicMovedRegionInfo(ServerNameserverName, longcloseSeqNum) @@ -271,7 +271,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html getServerName -publicServerNamegetServerName() +publicServerNamegetServerName() @@ -280,7 +280,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html getSeqNum -publiclonggetSeqNum() +publiclonggetSeqNum() @@ -289,7 +289,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html getMoveTime -publiclonggetMoveTime() +publiclonggetMoveTime() http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html index 2a00361..fa924cd 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -protected static final class HRegionServer.MovedRegionsCleaner +protected static final class HRegionServer.MovedRegionsCleaner extends ScheduledChore implements Stoppable Creates a Chore thread to clean the moved region cache. @@ -242,7 +242,7 @@ implements regionServer -privateHRegionServer regionServer +privateHRegionServer regionServer @@ -251,7 +251,7 @@ implements
[27/51] [partial] hbase-site git commit: Published site at 8bf966c8e936dec4d83bcbe85c5aab543f14a0df.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/27555316/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html index ea05301..26a93dd 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html @@ -269,3590 +269,3574 @@ 261 */ 262 protected ClusterConnection clusterConnection; 263 -264 /* -265 * Long-living meta table locator, which is created when the server is started and stopped -266 * when server shuts down. References to this locator shall be used to perform according -267 * operations in EventHandlers. Primary reason for this decision is to make it mockable -268 * for tests. -269 */ -270 protected MetaTableLocator metaTableLocator; -271 -272 /** -273 * Go here to get table descriptors. -274 */ -275 protected TableDescriptors tableDescriptors; -276 -277 // Replication services. If no replication, this handler will be null. -278 protected ReplicationSourceService replicationSourceHandler; -279 protected ReplicationSinkService replicationSinkHandler; -280 -281 // Compactions -282 public CompactSplit compactSplitThread; -283 -284 /** -285 * Map of regions currently being served by this region server. Key is the -286 * encoded region name. All access should be synchronized. -287 */ -288 protected final MapString, HRegion onlineRegions = new ConcurrentHashMap(); -289 -290 /** -291 * Map of encoded region names to the DataNode locations they should be hosted on -292 * We store the value as InetSocketAddress since this is used only in HDFS -293 * API (create() that takes favored nodes as hints for placing file blocks). -294 * We could have used ServerName here as the value class, but we'd need to -295 * convert it to InetSocketAddress at some point before the HDFS API call, and -296 * it seems a bit weird to store ServerName since ServerName refers to RegionServers -297 * and here we really mean DataNode locations. -298 */ -299 protected final MapString, InetSocketAddress[] regionFavoredNodesMap = -300 new ConcurrentHashMap(); -301 -302 // Leases -303 protected Leases leases; +264 /** +265 * Go here to get table descriptors. +266 */ +267 protected TableDescriptors tableDescriptors; +268 +269 // Replication services. If no replication, this handler will be null. +270 protected ReplicationSourceService replicationSourceHandler; +271 protected ReplicationSinkService replicationSinkHandler; +272 +273 // Compactions +274 public CompactSplit compactSplitThread; +275 +276 /** +277 * Map of regions currently being served by this region server. Key is the +278 * encoded region name. All access should be synchronized. +279 */ +280 protected final MapString, HRegion onlineRegions = new ConcurrentHashMap(); +281 +282 /** +283 * Map of encoded region names to the DataNode locations they should be hosted on +284 * We store the value as InetSocketAddress since this is used only in HDFS +285 * API (create() that takes favored nodes as hints for placing file blocks). +286 * We could have used ServerName here as the value class, but we'd need to +287 * convert it to InetSocketAddress at some point before the HDFS API call, and +288 * it seems a bit weird to store ServerName since ServerName refers to RegionServers +289 * and here we really mean DataNode locations. +290 */ +291 protected final MapString, InetSocketAddress[] regionFavoredNodesMap = +292 new ConcurrentHashMap(); +293 +294 // Leases +295 protected Leases leases; +296 +297 // Instance of the hbase executor executorService. +298 protected ExecutorService executorService; +299 +300 // If false, the file system has become unavailable +301 protected volatile boolean fsOk; +302 protected HFileSystem fs; +303 protected HFileSystem walFs; 304 -305 // Instance of the hbase executor executorService. -306 protected ExecutorService executorService; -307 -308 // If false, the file system has become unavailable -309 protected volatile boolean fsOk; -310 protected HFileSystem fs; -311 protected HFileSystem walFs; -312 -313 // Set when a report to the master comes back with a message asking us to -314 // shutdown. Also set by call to stop when debugging or running unit tests -315 // of HRegionServer in isolation. -316 private volatile boolean stopped = false; -317 -318 // Go down hard. Used if file system becomes unavailable and also in -319 // debugging and unit tests. -320 private volatile boolean abortRequested; -321 public static final String
[27/51] [partial] hbase-site git commit: Published site at 1acbd36c903b048141866b143507bfce124a5c5f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.ServerReportEvent.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.ServerReportEvent.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.ServerReportEvent.html deleted file mode 100644 index 908d0ce..000 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.ServerReportEvent.html +++ /dev/null @@ -1,165 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd;> - - - - - -Uses of Class org.apache.hadoop.hbase.master.assignment.ServerStateNode.ServerReportEvent (Apache HBase 3.0.0-SNAPSHOT API) - - - - - - - -JavaScript is disabled on your browser. - - - - - -Skip navigation links - - - - -Overview -Package -Class -Use -Tree -Deprecated -Index -Help - - - - -Prev -Next - - -Frames -NoFrames - - -AllClasses - - - - - - - - - - -Uses of Classorg.apache.hadoop.hbase.master.assignment.ServerStateNode.ServerReportEvent - - - - - -Packages that use ServerStateNode.ServerReportEvent - -Package -Description - - - -org.apache.hadoop.hbase.master.assignment - - - - - - - - - - -Uses of ServerStateNode.ServerReportEvent in org.apache.hadoop.hbase.master.assignment - -Fields in org.apache.hadoop.hbase.master.assignment declared as ServerStateNode.ServerReportEvent - -Modifier and Type -Field and Description - - - -private ServerStateNode.ServerReportEvent -ServerStateNode.reportEvent - - - - - - - - - - - - -Skip navigation links - - - - -Overview -Package -Class -Use -Tree -Deprecated -Index -Help - - - - -Prev -Next - - -Frames -NoFrames - - -AllClasses - - - - - - - - - -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - - http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.html index 44fe1fc..7e57352 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/ServerStateNode.html @@ -159,10 +159,6 @@ org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition.TransitionCodestate, longseqId) - -protected void -AssignmentManager.wakeServerReportEvent(ServerStateNodeserverNode) - http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/assignment/package-frame.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-frame.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-frame.html index 34c6cb8..03103e1 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-frame.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-frame.html @@ -37,7 +37,6 @@ RegionStateStore RegionTransitionProcedure ServerStateNode -ServerStateNode.ServerReportEvent SplitTableRegionProcedure TransitRegionStateProcedure UnassignProcedure http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html index 3235ced..577da39 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html @@ -210,22 +210,18 @@ -ServerStateNode.ServerReportEvent - - - SplitTableRegionProcedure The procedure to split a region in a table. -
[27/51] [partial] hbase-site git commit: Published site at 130057f13774f6b213cdb06952c805a29d59396e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/68eae623/devapidocs/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.html b/devapidocs/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.html index 44628e7..18eaf1f 100644 --- a/devapidocs/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.html +++ b/devapidocs/org/apache/hadoop/hbase/snapshot/RestoreSnapshotHelper.html @@ -110,7 +110,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public class RestoreSnapshotHelper +public class RestoreSnapshotHelper extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object Helper to Restore/Clone a Snapshot @@ -467,7 +467,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html LOG -private static finalorg.slf4j.Logger LOG +private static finalorg.slf4j.Logger LOG @@ -476,7 +476,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html regionsMap -private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Mapbyte[],byte[] regionsMap +private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Mapbyte[],byte[] regionsMap @@ -485,7 +485,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html parentsMap -private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,Pairhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String parentsMap +private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,Pairhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String parentsMap @@ -494,7 +494,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html monitor -private finalForeignExceptionDispatcher monitor +private finalForeignExceptionDispatcher monitor @@ -503,7 +503,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html status -private finalMonitoredTask status +private finalMonitoredTask status @@ -512,7 +512,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html snapshotManifest -private finalSnapshotManifest snapshotManifest +private finalSnapshotManifest snapshotManifest @@ -521,7 +521,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html snapshotDesc -private finalorg.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshotDesc +private finalorg.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshotDesc @@ -530,7 +530,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html snapshotTable -private finalTableName snapshotTable +private finalTableName snapshotTable @@ -539,7 +539,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html tableDesc -private finalTableDescriptor tableDesc +private finalTableDescriptor tableDesc @@ -548,7 +548,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html rootDir -private finalorg.apache.hadoop.fs.Path rootDir +private finalorg.apache.hadoop.fs.Path rootDir @@ -557,7 +557,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html tableDir -private finalorg.apache.hadoop.fs.Path tableDir +private finalorg.apache.hadoop.fs.Path tableDir @@ -566,7 +566,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html conf -private finalorg.apache.hadoop.conf.Configuration conf +private finalorg.apache.hadoop.conf.Configuration conf @@ -575,7 +575,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html fs -private finalorg.apache.hadoop.fs.FileSystem fs +private finalorg.apache.hadoop.fs.FileSystem fs @@ -584,7 +584,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
[27/51] [partial] hbase-site git commit: Published site at d5e4faacc354c1bc4d93efa71ca97ee3a056123e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b5e107c3/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html index 0af8acd..c5f21ac 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html @@ -645,1615 +645,1597 @@ 637 proc.afterReplay(getEnvironment()); 638 } 639}); -640 -641// 4. Push the procedures to the timeout executor -642waitingTimeoutList.forEach(proc - { -643 proc.afterReplay(getEnvironment()); -644 timeoutExecutor.add(proc); -645}); -646// 5. restore locks -647restoreLocks(); -648// 6. Push the procedure to the scheduler -649 failedList.forEach(scheduler::addBack); -650runnableList.forEach(p - { -651 p.afterReplay(getEnvironment()); -652 if (!p.hasParent()) { -653 sendProcedureLoadedNotification(p.getProcId()); -654 } -655 // If the procedure holds the lock, put the procedure in front -656 // If its parent holds the lock, put the procedure in front -657 // TODO. Is that possible that its ancestor holds the lock? -658 // For now, the deepest procedure hierarchy is: -659 // ModifyTableProcedure - ReopenTableProcedure - -660 // MoveTableProcedure - Unassign/AssignProcedure -661 // But ModifyTableProcedure and ReopenTableProcedure won't hold the lock -662 // So, check parent lock is enough(a tricky case is resovled by HBASE-21384). -663 // If some one change or add new procedures making 'grandpa' procedure -664 // holds the lock, but parent procedure don't hold the lock, there will -665 // be a problem here. We have to check one procedure's ancestors. -666 // And we need to change LockAndQueue.hasParentLock(Procedure? proc) method -667 // to check all ancestors too. -668 if (p.isLockedWhenLoading() || (p.hasParent() procedures -669 .get(p.getParentProcId()).isLockedWhenLoading())) { -670scheduler.addFront(p, false); -671 } else { -672// if it was not, it can wait. -673scheduler.addBack(p, false); -674 } -675}); -676// After all procedures put into the queue, signal the worker threads. -677// Otherwise, there is a race condition. See HBASE-21364. -678scheduler.signalAll(); -679 } +640// 4. restore locks +641restoreLocks(); +642 +643// 5. Push the procedures to the timeout executor +644waitingTimeoutList.forEach(proc - { +645 proc.afterReplay(getEnvironment()); +646 timeoutExecutor.add(proc); +647}); +648 +649// 6. Push the procedure to the scheduler +650 failedList.forEach(scheduler::addBack); +651runnableList.forEach(p - { +652 p.afterReplay(getEnvironment()); +653 if (!p.hasParent()) { +654 sendProcedureLoadedNotification(p.getProcId()); +655 } +656 scheduler.addBack(p); +657}); +658// After all procedures put into the queue, signal the worker threads. +659// Otherwise, there is a race condition. See HBASE-21364. +660scheduler.signalAll(); +661 } +662 +663 /** +664 * Initialize the procedure executor, but do not start workers. We will start them later. +665 * p/ +666 * It calls ProcedureStore.recoverLease() and ProcedureStore.load() to recover the lease, and +667 * ensure a single executor, and start the procedure replay to resume and recover the previous +668 * pending and in-progress procedures. +669 * @param numThreads number of threads available for procedure execution. +670 * @param abortOnCorruption true if you want to abort your service in case a corrupted procedure +671 * is found on replay. otherwise false. +672 */ +673 public void init(int numThreads, boolean abortOnCorruption) throws IOException { +674// We have numThreads executor + one timer thread used for timing out +675// procedures and triggering periodic procedures. +676this.corePoolSize = numThreads; +677this.maxPoolSize = 10 * numThreads; +678LOG.info("Starting {} core workers (bigger of cpus/4 or 16) with max (burst) worker count={}", +679corePoolSize, maxPoolSize); 680 -681 /** -682 * Initialize the procedure executor, but do not start workers. We will start them later. -683 * p/ -684 * It calls ProcedureStore.recoverLease() and ProcedureStore.load() to recover the lease, and -685 * ensure a single executor, and start the procedure replay to resume and recover the previous -686 * pending and in-progress procedures. -687 * @param numThreads number of threads available for procedure
[27/51] [partial] hbase-site git commit: Published site at 3fe8649b2c9ba1271c25e8f476548907e4c7a90d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8f09a71d/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html index c7d99b2..9d1542c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html @@ -382,1357 +382,1365 @@ 374for (int i = 0; i this.curFunctionCosts.length; i++) { 375 curFunctionCosts[i] = tempFunctionCosts[i]; 376} -377LOG.info("start StochasticLoadBalancer.balancer, initCost=" + currentCost + ", functionCost=" -378+ functionCost()); +377double initCost = currentCost; +378double newCost = currentCost; 379 -380double initCost = currentCost; -381double newCost = currentCost; -382 -383long computedMaxSteps; -384if (runMaxSteps) { -385 computedMaxSteps = Math.max(this.maxSteps, -386 ((long)cluster.numRegions * (long)this.stepsPerRegion * (long)cluster.numServers)); -387} else { -388 computedMaxSteps = Math.min(this.maxSteps, -389 ((long)cluster.numRegions * (long)this.stepsPerRegion * (long)cluster.numServers)); -390} -391// Perform a stochastic walk to see if we can get a good fit. -392long step; -393 -394for (step = 0; step computedMaxSteps; step++) { -395 Cluster.Action action = nextAction(cluster); -396 -397 if (action.type == Type.NULL) { -398continue; -399 } -400 -401 cluster.doAction(action); -402 updateCostsWithAction(cluster, action); -403 -404 newCost = computeCost(cluster, currentCost); -405 -406 // Should this be kept? -407 if (newCost currentCost) { -408currentCost = newCost; -409 -410// save for JMX -411curOverallCost = currentCost; -412for (int i = 0; i this.curFunctionCosts.length; i++) { -413 curFunctionCosts[i] = tempFunctionCosts[i]; -414} -415 } else { -416// Put things back the way they were before. -417// TODO: undo by remembering old values -418Action undoAction = action.undoAction(); -419cluster.doAction(undoAction); -420updateCostsWithAction(cluster, undoAction); -421 } -422 -423 if (EnvironmentEdgeManager.currentTime() - startTime -424 maxRunningTime) { -425break; -426 } -427} -428long endTime = EnvironmentEdgeManager.currentTime(); -429 -430 metricsBalancer.balanceCluster(endTime - startTime); -431 -432// update costs metrics -433updateStochasticCosts(tableName, curOverallCost, curFunctionCosts); -434if (initCost currentCost) { -435 plans = createRegionPlans(cluster); -436 LOG.info("Finished computing new load balance plan. Computation took {}" + -437" to try {} different iterations. Found a solution that moves " + -438"{} regions; Going from a computed cost of {}" + -439" to a new cost of {}", java.time.Duration.ofMillis(endTime - startTime), -440step, plans.size(), initCost, currentCost); -441 return plans; -442} -443LOG.info("Could not find a better load balance plan. Tried {} different configurations in " + -444 "{}, and did not find anything with a computed cost less than {}", step, -445 java.time.Duration.ofMillis(endTime - startTime), initCost); -446return null; -447 } -448 -449 /** -450 * update costs to JMX -451 */ -452 private void updateStochasticCosts(TableName tableName, Double overall, Double[] subCosts) { -453if (tableName == null) return; -454 -455// check if the metricsBalancer is MetricsStochasticBalancer before casting -456if (metricsBalancer instanceof MetricsStochasticBalancer) { -457 MetricsStochasticBalancer balancer = (MetricsStochasticBalancer) metricsBalancer; -458 // overall cost -459 balancer.updateStochasticCost(tableName.getNameAsString(), -460"Overall", "Overall cost", overall); -461 -462 // each cost function -463 for (int i = 0; i costFunctions.length; i++) { -464CostFunction costFunction = costFunctions[i]; -465String costFunctionName = costFunction.getClass().getSimpleName(); -466Double costPercent = (overall == 0) ? 0 : (subCosts[i] / overall); -467// TODO: cost function may need a specific description -468 balancer.updateStochasticCost(tableName.getNameAsString(), costFunctionName, -469 "The percent of " + costFunctionName, costPercent); -470 } -471} -472 } -473 -474 private String
[27/51] [partial] hbase-site git commit: Published site at 7adf590106826b9e4432cfeee06acdc0ccff8c6e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/425db230/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureTree.Entry.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureTree.Entry.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureTree.Entry.html new file mode 100644 index 000..8ef0ff5 --- /dev/null +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureTree.Entry.html @@ -0,0 +1,371 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + +Source code + + + + +001/** +002 * Licensed to the Apache Software Foundation (ASF) under one +003 * or more contributor license agreements. See the NOTICE file +004 * distributed with this work for additional information +005 * regarding copyright ownership. The ASF licenses this file +006 * to you under the Apache License, Version 2.0 (the +007 * "License"); you may not use this file except in compliance +008 * with the License. You may obtain a copy of the License at +009 * +010 * http://www.apache.org/licenses/LICENSE-2.0 +011 * +012 * Unless required by applicable law or agreed to in writing, software +013 * distributed under the License is distributed on an "AS IS" BASIS, +014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +015 * See the License for the specific language governing permissions and +016 * limitations under the License. +017 */ +018package org.apache.hadoop.hbase.procedure2.store.wal; +019 +020import java.io.IOException; +021import java.util.ArrayList; +022import java.util.Collection; +023import java.util.Collections; +024import java.util.Comparator; +025import java.util.HashMap; +026import java.util.Iterator; +027import java.util.List; +028import java.util.Map; +029import java.util.NoSuchElementException; +030import org.apache.commons.lang3.mutable.MutableInt; +031import org.apache.hadoop.hbase.procedure2.Procedure; +032import org.apache.hadoop.hbase.procedure2.ProcedureUtil; +033import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator; +034import org.apache.yetus.audience.InterfaceAudience; +035import org.slf4j.Logger; +036import org.slf4j.LoggerFactory; +037 +038import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos; +039 +040/** +041 * Used to build the tree for procedures. +042 * p/ +043 * We will group the procedures with the root procedure, and then validate each group. For each +044 * group of procedures(with the same root procedure), we will collect all the stack ids, if the max +045 * stack id is n, then all the stack ids should be from 0 to n, non-repetition and non-omission. If +046 * not, we will consider all the procedures in this group as corrupted. Please see the code in +047 * {@link #checkReady(Entry, Map)} method. +048 * p/ +049 * For the procedures not in any group, i.e, can not find the root procedure for these procedures, +050 * we will also consider them as corrupted. Please see the code in {@link #checkOrphan(Map)} method. +051 */ +052@InterfaceAudience.Private +053public final class WALProcedureTree { +054 +055 private static final Logger LOG = LoggerFactory.getLogger(WALProcedureTree.class); +056 +057 private static final class Entry { +058 +059private final ProcedureProtos.Procedure proc; +060 +061private final ListEntry subProcs = new ArrayList(); +062 +063public Entry(ProcedureProtos.Procedure proc) { +064 this.proc = proc; +065} +066 +067@Override +068public String toString() { +069 StringBuilder sb = new StringBuilder(); +070 sb.append("Procedure(pid="); +071 sb.append(proc.getProcId()); +072 sb.append(", ppid="); +073 sb.append(proc.hasParentId() ? proc.getParentId() : Procedure.NO_PROC_ID); +074 sb.append(", class="); +075 sb.append(proc.getClassName()); +076 sb.append(")"); +077 return sb.toString(); +078} +079 } +080 +081 // when loading we will iterator the procedures twice, so use this class to cache the deserialized +082 // result to prevent deserializing multiple times. +083 private static final class ProtoAndProc { +084private final ProcedureProtos.Procedure proto; +085 +086private Procedure? proc; +087 +088public ProtoAndProc(ProcedureProtos.Procedure proto) { +089 this.proto = proto; +090} +091 +092public Procedure? getProc() throws IOException { +093 if (proc == null) { +094proc = ProcedureUtil.convertToProcedure(proto); +095 } +096 return proc; +097} +098 } +099 +100 private final ListProtoAndProc validProcs = new ArrayList(); +101 +102 private final ListProtoAndProc corruptedProcs = new ArrayList(); +103 +104 private static boolean isFinished(ProcedureProtos.Procedure proc) { +105if (!proc.hasParentId()) { +106 switch (proc.getState())
[27/51] [partial] hbase-site git commit: Published site at 5fbb227deb365fe812d433fe39b85ac4b0ddee20.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html b/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html index 3d27554..fd63f54 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html +++ b/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -static class StochasticLoadBalancer.RegionReplicaCandidateGenerator +static class StochasticLoadBalancer.RegionReplicaCandidateGenerator extends StochasticLoadBalancer.CandidateGenerator Generates candidates which moves the replicas out of the region server for co-hosted region replicas @@ -228,7 +228,7 @@ extends randomGenerator -StochasticLoadBalancer.RandomCandidateGenerator randomGenerator +StochasticLoadBalancer.RandomCandidateGenerator randomGenerator @@ -245,7 +245,7 @@ extends RegionReplicaCandidateGenerator -RegionReplicaCandidateGenerator() +RegionReplicaCandidateGenerator() @@ -262,7 +262,7 @@ extends selectCoHostedRegionPerGroup -intselectCoHostedRegionPerGroup(int[]primariesOfRegionsPerGroup, +intselectCoHostedRegionPerGroup(int[]primariesOfRegionsPerGroup, int[]regionsPerGroup, int[]regionIndexToPrimaryIndex) Randomly select one regionIndex out of all region replicas co-hosted in the same group @@ -284,7 +284,7 @@ extends generate -BaseLoadBalancer.Cluster.Actiongenerate(BaseLoadBalancer.Clustercluster) +BaseLoadBalancer.Cluster.Actiongenerate(BaseLoadBalancer.Clustercluster) Specified by: generatein classStochasticLoadBalancer.CandidateGenerator http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html b/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html index 1faf326..76e8473 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html +++ b/devapidocs/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -static class StochasticLoadBalancer.RegionReplicaHostCostFunction +static class StochasticLoadBalancer.RegionReplicaHostCostFunction extends StochasticLoadBalancer.CostFunction A cost function for region replicas. We give a very high cost to hosting replicas of the same region in the same host. We do not prevent the case @@ -272,7 +272,7 @@ extends REGION_REPLICA_HOST_COST_KEY -private static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String REGION_REPLICA_HOST_COST_KEY +private static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String REGION_REPLICA_HOST_COST_KEY See Also: Constant Field Values @@ -285,7 +285,7 @@ extends DEFAULT_REGION_REPLICA_HOST_COST_KEY -private static finalfloat DEFAULT_REGION_REPLICA_HOST_COST_KEY +private static finalfloat DEFAULT_REGION_REPLICA_HOST_COST_KEY See Also: Constant Field Values @@ -298,7 +298,7 @@ extends maxCost -long maxCost +long maxCost @@ -307,7 +307,7 @@ extends costsPerGroup -long[] costsPerGroup +long[] costsPerGroup @@ -316,7 +316,7 @@ extends primariesOfRegionsPerGroup -int[][] primariesOfRegionsPerGroup +int[][] primariesOfRegionsPerGroup @@ -333,7 +333,7 @@ extends RegionReplicaHostCostFunction -publicRegionReplicaHostCostFunction(org.apache.hadoop.conf.Configurationconf) +publicRegionReplicaHostCostFunction(org.apache.hadoop.conf.Configurationconf) @@ -350,7 +350,7 @@ extends init -voidinit(BaseLoadBalancer.Clustercluster) +voidinit(BaseLoadBalancer.Clustercluster) Description copied from class:StochasticLoadBalancer.CostFunction Called once per LB invocation to give the cost function to initialize it's state, and perform any costly calculation. @@ -366,7 +366,7 @@ extends getMaxCost -longgetMaxCost(BaseLoadBalancer.Clustercluster) +longgetMaxCost(BaseLoadBalancer.Clustercluster) @@ -375,7 +375,7 @@ extends isNeeded -booleanisNeeded()
[27/51] [partial] hbase-site git commit: Published site at 821e4d7de2d576189f4288d1c2acf9e9a9471f5c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/323b17d9/testdevapidocs/org/apache/hadoop/hbase/backup/package-tree.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/backup/package-tree.html b/testdevapidocs/org/apache/hadoop/hbase/backup/package-tree.html index 2d9b60f..6af42fe 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/backup/package-tree.html +++ b/testdevapidocs/org/apache/hadoop/hbase/backup/package-tree.html @@ -146,8 +146,8 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.backup.TestBackupDeleteWithFailures.Failure org.apache.hadoop.hbase.backup.TestIncrementalBackupMergeWithFailures.FailurePhase +org.apache.hadoop.hbase.backup.TestBackupDeleteWithFailures.Failure http://git-wip-us.apache.org/repos/asf/hbase-site/blob/323b17d9/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html index 06d362a..cb399cd 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html +++ b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html @@ -1167,8 +1167,16 @@ static HBaseClassTestRule +TestRestoreSnapshotFromClientGetCompactionState.CLASS_RULE + + +static HBaseClassTestRule TestMobCloneSnapshotFromClientNormal.CLASS_RULE + +static HBaseClassTestRule +TestMobRestoreSnapshotFromClientClone.CLASS_RULE + static HBaseClassTestRule TestCISleep.CLASS_RULE @@ -1223,43 +1231,43 @@ static HBaseClassTestRule -TestAsyncTableGetMultiThreaded.CLASS_RULE +TestMobRestoreSnapshotFromClientAfterSplittingRegions.CLASS_RULE static HBaseClassTestRule -TestSnapshotCloneIndependence.CLASS_RULE +TestAsyncTableGetMultiThreaded.CLASS_RULE static HBaseClassTestRule -TestCIDeleteRpcTimeout.CLASS_RULE +TestSnapshotCloneIndependence.CLASS_RULE static HBaseClassTestRule -TestZKAsyncRegistry.CLASS_RULE +TestCIDeleteRpcTimeout.CLASS_RULE static HBaseClassTestRule -TestAsyncProcedureAdminApi.CLASS_RULE +TestZKAsyncRegistry.CLASS_RULE static HBaseClassTestRule -TestAsyncClusterAdminApi2.CLASS_RULE +TestAsyncProcedureAdminApi.CLASS_RULE static HBaseClassTestRule -TestCIGetOperationTimeout.CLASS_RULE +TestAsyncClusterAdminApi2.CLASS_RULE static HBaseClassTestRule -TestAlwaysSetScannerId.CLASS_RULE +TestCIGetOperationTimeout.CLASS_RULE static HBaseClassTestRule -TestSnapshotFromClient.CLASS_RULE +TestAlwaysSetScannerId.CLASS_RULE static HBaseClassTestRule -TestRestoreSnapshotFromClient.CLASS_RULE +TestSnapshotFromClient.CLASS_RULE static HBaseClassTestRule @@ -1311,56 +1319,64 @@ static HBaseClassTestRule -TestBlockEvictionFromClient.CLASS_RULE +TestMobRestoreSnapshotFromClientGetCompactionState.CLASS_RULE static HBaseClassTestRule -TestCIBadHostname.CLASS_RULE +TestBlockEvictionFromClient.CLASS_RULE static HBaseClassTestRule -TestCloneSnapshotFromClientCloneLinksAfterDelete.CLASS_RULE +TestCIBadHostname.CLASS_RULE static HBaseClassTestRule -TestAsyncTableAdminApi2.CLASS_RULE +TestCloneSnapshotFromClientCloneLinksAfterDelete.CLASS_RULE static HBaseClassTestRule -TestAsyncQuotaAdminApi.CLASS_RULE +TestAsyncTableAdminApi2.CLASS_RULE static HBaseClassTestRule -TestAsyncMetaRegionLocator.CLASS_RULE +TestAsyncQuotaAdminApi.CLASS_RULE static HBaseClassTestRule -TestAsyncTableNoncedRetry.CLASS_RULE +TestAsyncMetaRegionLocator.CLASS_RULE static HBaseClassTestRule -TestRegionLocationCaching.CLASS_RULE +TestAsyncTableNoncedRetry.CLASS_RULE static HBaseClassTestRule -TestConnectionUtils.CLASS_RULE +TestRegionLocationCaching.CLASS_RULE static HBaseClassTestRule -TestMobCloneSnapshotFromClientCloneLinksAfterDelete.CLASS_RULE +TestConnectionUtils.CLASS_RULE static HBaseClassTestRule -TestAsyncTableScanMetrics.CLASS_RULE +TestMobCloneSnapshotFromClientCloneLinksAfterDelete.CLASS_RULE static HBaseClassTestRule -TestFromClientSideScanExcpetionWithCoprocessor.CLASS_RULE +TestAsyncTableScanMetrics.CLASS_RULE static HBaseClassTestRule +TestFromClientSideScanExcpetionWithCoprocessor.CLASS_RULE + + +static HBaseClassTestRule TestMobCloneSnapshotFromClientError.CLASS_RULE + +static HBaseClassTestRule +TestRestoreSnapshotFromClientSchemaChange.CLASS_RULE + static HBaseClassTestRule TestAsyncDecommissionAdminApi.CLASS_RULE @@
[27/51] [partial] hbase-site git commit: Published site at fa5fa6ecdd071b72b58971058ff3ab9d28c3e709.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d1341859/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.ProcedureExecutorListener.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.ProcedureExecutorListener.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.ProcedureExecutorListener.html index 061ce80..bdfc3f8 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.ProcedureExecutorListener.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.ProcedureExecutorListener.html @@ -39,2126 +39,2163 @@ 031import java.util.Set; 032import java.util.concurrent.ConcurrentHashMap; 033import java.util.concurrent.CopyOnWriteArrayList; -034import java.util.concurrent.TimeUnit; -035import java.util.concurrent.atomic.AtomicBoolean; -036import java.util.concurrent.atomic.AtomicInteger; -037import java.util.concurrent.atomic.AtomicLong; -038import java.util.stream.Collectors; -039import java.util.stream.Stream; -040import org.apache.hadoop.conf.Configuration; -041import org.apache.hadoop.hbase.HConstants; -042import org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException; -043import org.apache.hadoop.hbase.log.HBaseMarkers; -044import org.apache.hadoop.hbase.procedure2.Procedure.LockState; -045import org.apache.hadoop.hbase.procedure2.store.ProcedureStore; -046import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator; -047import org.apache.hadoop.hbase.procedure2.util.StringUtils; -048import org.apache.hadoop.hbase.security.User; -049import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -050import org.apache.hadoop.hbase.util.IdLock; -051import org.apache.hadoop.hbase.util.NonceKey; -052import org.apache.hadoop.hbase.util.Threads; -053import org.apache.yetus.audience.InterfaceAudience; -054import org.slf4j.Logger; -055import org.slf4j.LoggerFactory; -056 -057import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; -058import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; +034import java.util.concurrent.Executor; +035import java.util.concurrent.Executors; +036import java.util.concurrent.TimeUnit; +037import java.util.concurrent.atomic.AtomicBoolean; +038import java.util.concurrent.atomic.AtomicInteger; +039import java.util.concurrent.atomic.AtomicLong; +040import java.util.stream.Collectors; +041import java.util.stream.Stream; +042import org.apache.hadoop.conf.Configuration; +043import org.apache.hadoop.hbase.HConstants; +044import org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException; +045import org.apache.hadoop.hbase.log.HBaseMarkers; +046import org.apache.hadoop.hbase.procedure2.Procedure.LockState; +047import org.apache.hadoop.hbase.procedure2.store.ProcedureStore; +048import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator; +049import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureStoreListener; +050import org.apache.hadoop.hbase.procedure2.util.StringUtils; +051import org.apache.hadoop.hbase.security.User; +052import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; +053import org.apache.hadoop.hbase.util.IdLock; +054import org.apache.hadoop.hbase.util.NonceKey; +055import org.apache.hadoop.hbase.util.Threads; +056import org.apache.yetus.audience.InterfaceAudience; +057import org.slf4j.Logger; +058import org.slf4j.LoggerFactory; 059 -060import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState; -061 -062/** -063 * Thread Pool that executes the submitted procedures. -064 * The executor has a ProcedureStore associated. -065 * Each operation is logged and on restart the pending procedures are resumed. -066 * -067 * Unless the Procedure code throws an error (e.g. invalid user input) -068 * the procedure will complete (at some point in time), On restart the pending -069 * procedures are resumed and the once failed will be rolledback. +060import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; +061import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; +062import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; +063 +064import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState; +065 +066/** +067 * Thread Pool that executes the submitted procedures. +068 * The executor has a ProcedureStore associated. +069 * Each operation is logged and on restart the pending procedures are resumed. 070 * -071 * The user can add procedures to the executor via submitProcedure(proc) -072 * check for the finished state via isFinished(procId) -073 * and get the result via getResult(procId) -074 */ -075@InterfaceAudience.Private -076public class ProcedureExecutorTEnvironment {
[27/51] [partial] hbase-site git commit: Published site at 6bc7089f9e0793efc9bdd46a84f5ccd9bc4579ad.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/419d0338/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html index c7f1e58..c74635e 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html @@ -164,897 +164,923 @@ 156 */ 157 private volatile boolean bypass = false; 158 -159 public boolean isBypass() { -160return bypass; -161 } -162 -163 /** -164 * set the bypass to true -165 * Only called in {@link ProcedureExecutor#bypassProcedure(long, long, boolean)} for now, -166 * DO NOT use this method alone, since we can't just bypass -167 * one single procedure. We need to bypass its ancestor too. So making it package private -168 */ -169 void bypass() { -170this.bypass = true; -171 } +159 /** +160 * Indicate whether we need to persist the procedure to ProcedureStore after execution. Default to +161 * true, and the implementation can all {@link #skipPersistence()} to let the framework skip the +162 * persistence of the procedure. +163 * p/ +164 * This is useful when the procedure is in error and you want to retry later. The retry interval +165 * and the number of retries are usually not critical so skip the persistence can save some +166 * resources, and also speed up the restart processing. +167 * p/ +168 * Notice that this value will be reset to true every time before execution. And when rolling back +169 * we do not test this value. +170 */ +171 private boolean persist = true; 172 -173 /** -174 * The main code of the procedure. It must be idempotent since execute() -175 * may be called multiple times in case of machine failure in the middle -176 * of the execution. -177 * @param env the environment passed to the ProcedureExecutor -178 * @return a set of sub-procedures to run or ourselves if there is more work to do or null if the -179 * procedure is done. -180 * @throws ProcedureYieldException the procedure will be added back to the queue and retried later. -181 * @throws InterruptedException the procedure will be added back to the queue and retried later. -182 * @throws ProcedureSuspendedException Signal to the executor that Procedure has suspended itself and -183 * has set itself up waiting for an external event to wake it back up again. -184 */ -185 protected abstract ProcedureTEnvironment[] execute(TEnvironment env) -186throws ProcedureYieldException, ProcedureSuspendedException, InterruptedException; -187 -188 /** -189 * The code to undo what was done by the execute() code. -190 * It is called when the procedure or one of the sub-procedures failed or an -191 * abort was requested. It should cleanup all the resources created by -192 * the execute() call. The implementation must be idempotent since rollback() -193 * may be called multiple time in case of machine failure in the middle -194 * of the execution. -195 * @param env the environment passed to the ProcedureExecutor -196 * @throws IOException temporary failure, the rollback will retry later -197 * @throws InterruptedException the procedure will be added back to the queue and retried later -198 */ -199 protected abstract void rollback(TEnvironment env) -200throws IOException, InterruptedException; -201 -202 /** -203 * The abort() call is asynchronous and each procedure must decide how to deal -204 * with it, if they want to be abortable. The simplest implementation -205 * is to have an AtomicBoolean set in the abort() method and then the execute() -206 * will check if the abort flag is set or not. -207 * abort() may be called multiple times from the client, so the implementation -208 * must be idempotent. -209 * -210 * pNOTE: abort() is not like Thread.interrupt(). It is just a notification -211 * that allows the procedure implementor abort. -212 */ -213 protected abstract boolean abort(TEnvironment env); -214 -215 /** -216 * The user-level code of the procedure may have some state to -217 * persist (e.g. input arguments or current position in the processing state) to -218 * be able to resume on failure. -219 * @param serializer stores the serializable state -220 */ -221 protected abstract void serializeStateData(ProcedureStateSerializer serializer) -222throws IOException; -223 -224 /** -225 * Called on store load to allow the user to decode the previously serialized -226 * state. -227 * @param serializer contains the serialized state -228 */ -229 protected abstract void deserializeStateData(ProcedureStateSerializer serializer) -230throws IOException; -231 -232 /** -233 * The {@link
[27/51] [partial] hbase-site git commit: Published site at d7e08317d2f214e4cca7b67578aba0ed7a567d54.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegionFileSystem.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegionFileSystem.html b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegionFileSystem.html index 55d677f..aae23e1 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegionFileSystem.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/HRegionFileSystem.html @@ -212,6 +212,10 @@ HStore.getRegionFileSystem() +(package private) HRegionFileSystem +HRegion.getRegionWALFileSystem() + + static HRegionFileSystem HRegionFileSystem.openRegionFromFileSystem(org.apache.hadoop.conf.Configurationconf, org.apache.hadoop.fs.FileSystemfs, http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/LastSequenceId.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/LastSequenceId.html b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/LastSequenceId.html index 4c2b975..2abca97 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/LastSequenceId.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/LastSequenceId.html @@ -203,7 +203,7 @@ static boolean WALSplitter.splitLogFile(org.apache.hadoop.fs.PathwalDir, org.apache.hadoop.fs.FileStatuslogfile, -org.apache.hadoop.fs.FileSystemfs, +org.apache.hadoop.fs.FileSystemwalFS, org.apache.hadoop.conf.Configurationconf, CancelableProgressablereporter, LastSequenceIdidChecker, @@ -224,7 +224,7 @@ WALSplitter(WALFactoryfactory, org.apache.hadoop.conf.Configurationconf, org.apache.hadoop.fs.PathwalDir, - org.apache.hadoop.fs.FileSystemfs, + org.apache.hadoop.fs.FileSystemwalFS, LastSequenceIdidChecker, SplitLogWorkerCoordinationsplitLogWorkerCoordination) http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/org/apache/hadoop/hbase/regionserver/handler/AssignRegionHandler.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/handler/AssignRegionHandler.html b/devapidocs/org/apache/hadoop/hbase/regionserver/handler/AssignRegionHandler.html index 2ba4882..d9b1b13 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/handler/AssignRegionHandler.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/handler/AssignRegionHandler.html @@ -119,7 +119,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public class AssignRegionHandler +public class AssignRegionHandler extends EventHandler Handles opening of a region on a region server. @@ -157,6 +157,10 @@ extends regionInfo +private RetryCounter +retryCounter + + private TableDescriptor tableDesc @@ -266,7 +270,7 @@ extends LOG -private static finalorg.slf4j.Logger LOG +private static finalorg.slf4j.Logger LOG @@ -275,7 +279,7 @@ extends regionInfo -private finalRegionInfo regionInfo +private finalRegionInfo regionInfo @@ -284,16 +288,25 @@ extends tableDesc -private finalTableDescriptor tableDesc +private finalTableDescriptor tableDesc - + masterSystemTime -private finallong masterSystemTime +private finallong masterSystemTime + + + + + + + +retryCounter +private finalRetryCounter retryCounter @@ -310,7 +323,7 @@ extends AssignRegionHandler -publicAssignRegionHandler(RegionServerServicesserver, +publicAssignRegionHandler(RegionServerServicesserver, RegionInforegionInfo, @Nullable TableDescriptortableDesc, @@ -332,7 +345,7 @@ extends getServer -privateRegionServerServicesgetServer() +privateRegionServerServicesgetServer() @@ -341,7 +354,7 @@ extends cleanUpAndReportFailure -privatevoidcleanUpAndReportFailure(https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOExceptionerror) +privatevoidcleanUpAndReportFailure(https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOExceptionerror) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Throws: @@ -355,7 +368,7 @@ extends process -publicvoidprocess() +publicvoidprocess() throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in
[27/51] [partial] hbase-site git commit: Published site at 8eaaa63114a64bcaeaf0ed9bdd88615ee22255c1.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f6f9d4f3/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html b/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html index 49f081b..33c9cc0 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html @@ -35,309 +35,328 @@ 027import java.util.concurrent.BlockingQueue; 028import java.util.concurrent.ConcurrentHashMap; 029import java.util.concurrent.ConcurrentMap; -030import java.util.concurrent.LinkedBlockingQueue; -031import java.util.concurrent.ThreadPoolExecutor; -032import java.util.concurrent.TimeUnit; -033import java.util.concurrent.atomic.AtomicLong; -034 -035import org.apache.yetus.audience.InterfaceAudience; -036import org.slf4j.Logger; -037import org.slf4j.LoggerFactory; -038import org.apache.hadoop.hbase.monitoring.ThreadMonitoring; +030import java.util.concurrent.Executors; +031import java.util.concurrent.LinkedBlockingQueue; +032import java.util.concurrent.ThreadPoolExecutor; +033import java.util.concurrent.TimeUnit; +034import java.util.concurrent.atomic.AtomicLong; +035import org.apache.hadoop.hbase.monitoring.ThreadMonitoring; +036import org.apache.yetus.audience.InterfaceAudience; +037import org.slf4j.Logger; +038import org.slf4j.LoggerFactory; 039 040import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; 041import org.apache.hbase.thirdparty.com.google.common.collect.Lists; 042import org.apache.hbase.thirdparty.com.google.common.collect.Maps; -043import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -044 -045/** -046 * This is a generic executor service. This component abstracts a -047 * threadpool, a queue to which {@link EventType}s can be submitted, -048 * and a codeRunnable/code that handles the object that is added to the queue. -049 * -050 * pIn order to create a new service, create an instance of this class and -051 * then do: codeinstance.startExecutorService("myService");/code. When done -052 * call {@link #shutdown()}. -053 * -054 * pIn order to use the service created above, call -055 * {@link #submit(EventHandler)}. -056 */ -057@InterfaceAudience.Private -058public class ExecutorService { -059 private static final Logger LOG = LoggerFactory.getLogger(ExecutorService.class); -060 -061 // hold the all the executors created in a map addressable by their names -062 private final ConcurrentHashMapString, Executor executorMap = new ConcurrentHashMap(); +043import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ListenableFuture; +044import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ListeningScheduledExecutorService; +045import org.apache.hbase.thirdparty.com.google.common.util.concurrent.MoreExecutors; +046import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; +047 +048/** +049 * This is a generic executor service. This component abstracts a +050 * threadpool, a queue to which {@link EventType}s can be submitted, +051 * and a codeRunnable/code that handles the object that is added to the queue. +052 * +053 * pIn order to create a new service, create an instance of this class and +054 * then do: codeinstance.startExecutorService("myService");/code. When done +055 * call {@link #shutdown()}. +056 * +057 * pIn order to use the service created above, call +058 * {@link #submit(EventHandler)}. +059 */ +060@InterfaceAudience.Private +061public class ExecutorService { +062 private static final Logger LOG = LoggerFactory.getLogger(ExecutorService.class); 063 -064 // Name of the server hosting this executor service. -065 private final String servername; +064 // hold the all the executors created in a map addressable by their names +065 private final ConcurrentHashMapString, Executor executorMap = new ConcurrentHashMap(); 066 -067 /** -068 * Default constructor. -069 * @param servername Name of the hosting server. -070 */ -071 public ExecutorService(final String servername) { -072super(); -073this.servername = servername; -074 } -075 -076 /** -077 * Start an executor service with a given name. If there was a service already -078 * started with the same name, this throws a RuntimeException. -079 * @param name Name of the service to start. -080 */ -081 @VisibleForTesting -082 public void startExecutorService(String name, int maxThreads) { -083if (this.executorMap.get(name) != null) { -084 throw new RuntimeException("An executor service with the name " + name + -085" is already running!"); -086
[27/51] [partial] hbase-site git commit: Published site at cd161d976ef47b84e904f2d54bac65d2f3417c2a.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fa1bebf8/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ProcedureFuture.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ProcedureFuture.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ProcedureFuture.html index a5789e0..93a57cb 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ProcedureFuture.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ProcedureFuture.html @@ -238,4120 +238,4119 @@ 230 * @see Admin 231 */ 232@InterfaceAudience.Private -233@InterfaceStability.Evolving -234public class HBaseAdmin implements Admin { -235 private static final Logger LOG = LoggerFactory.getLogger(HBaseAdmin.class); -236 -237 private ClusterConnection connection; -238 -239 private final Configuration conf; -240 private final long pause; -241 private final int numRetries; -242 private final int syncWaitTimeout; -243 private boolean aborted; -244 private int operationTimeout; -245 private int rpcTimeout; -246 -247 private RpcRetryingCallerFactory rpcCallerFactory; -248 private RpcControllerFactory rpcControllerFactory; -249 -250 private NonceGenerator ng; -251 -252 @Override -253 public int getOperationTimeout() { -254return operationTimeout; -255 } -256 -257 HBaseAdmin(ClusterConnection connection) throws IOException { -258this.conf = connection.getConfiguration(); -259this.connection = connection; -260 -261// TODO: receive ConnectionConfiguration here rather than re-parsing these configs every time. -262this.pause = this.conf.getLong(HConstants.HBASE_CLIENT_PAUSE, -263 HConstants.DEFAULT_HBASE_CLIENT_PAUSE); -264this.numRetries = this.conf.getInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, -265 HConstants.DEFAULT_HBASE_CLIENT_RETRIES_NUMBER); -266this.operationTimeout = this.conf.getInt(HConstants.HBASE_CLIENT_OPERATION_TIMEOUT, -267 HConstants.DEFAULT_HBASE_CLIENT_OPERATION_TIMEOUT); -268this.rpcTimeout = this.conf.getInt(HConstants.HBASE_RPC_TIMEOUT_KEY, -269 HConstants.DEFAULT_HBASE_RPC_TIMEOUT); -270this.syncWaitTimeout = this.conf.getInt( -271 "hbase.client.sync.wait.timeout.msec", 10 * 6); // 10min -272 -273this.rpcCallerFactory = connection.getRpcRetryingCallerFactory(); -274this.rpcControllerFactory = connection.getRpcControllerFactory(); -275 -276this.ng = this.connection.getNonceGenerator(); -277 } -278 -279 @Override -280 public void abort(String why, Throwable e) { -281// Currently does nothing but throw the passed message and exception -282this.aborted = true; -283throw new RuntimeException(why, e); -284 } -285 -286 @Override -287 public boolean isAborted() { -288return this.aborted; -289 } -290 -291 @Override -292 public boolean abortProcedure(final long procId, final boolean mayInterruptIfRunning) -293 throws IOException { -294return get(abortProcedureAsync(procId, mayInterruptIfRunning), this.syncWaitTimeout, -295 TimeUnit.MILLISECONDS); -296 } -297 -298 @Override -299 public FutureBoolean abortProcedureAsync(final long procId, final boolean mayInterruptIfRunning) -300 throws IOException { -301Boolean abortProcResponse = -302executeCallable(new MasterCallableAbortProcedureResponse(getConnection(), -303getRpcControllerFactory()) { -304 @Override -305 protected AbortProcedureResponse rpcCall() throws Exception { -306AbortProcedureRequest abortProcRequest = -307 AbortProcedureRequest.newBuilder().setProcId(procId).build(); -308return master.abortProcedure(getRpcController(), abortProcRequest); -309 } -310}).getIsProcedureAborted(); -311return new AbortProcedureFuture(this, procId, abortProcResponse); -312 } -313 -314 @Override -315 public ListTableDescriptor listTableDescriptors() throws IOException { -316return listTableDescriptors((Pattern)null, false); -317 } -318 -319 @Override -320 public ListTableDescriptor listTableDescriptors(Pattern pattern) throws IOException { -321return listTableDescriptors(pattern, false); -322 } -323 -324 @Override -325 public ListTableDescriptor listTableDescriptors(Pattern pattern, boolean includeSysTables) -326 throws IOException { -327return executeCallable(new MasterCallableListTableDescriptor(getConnection(), -328getRpcControllerFactory()) { -329 @Override -330 protected ListTableDescriptor rpcCall() throws Exception { -331GetTableDescriptorsRequest req = -332 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables); -333return ProtobufUtil.toTableDescriptorList(master.getTableDescriptors(getRpcController(), -334req)); -335 } -336}); -337 } -338 -339 @Override -340
[27/51] [partial] hbase-site git commit: Published site at c6a65ba63fce85ac7c4b62b96ef2bbe6c35d2f00.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/293abb17/devapidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html b/devapidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html index 2036bc9..bcfe1b9 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html @@ -28,113 +28,124 @@ 020 021import java.io.IOException; 022import java.util.ArrayList; -023 -024import org.apache.hadoop.hbase.Cell; -025import org.apache.yetus.audience.InterfaceAudience; -026import org.apache.hadoop.hbase.exceptions.DeserializationException; -027import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos; -028 -029import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; -030import org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException; -031 -032/** -033 * A filter that will only return the first KV from each row. -034 * p -035 * This filter can be used to more efficiently perform row count operations. -036 */ -037@InterfaceAudience.Public -038public class FirstKeyOnlyFilter extends FilterBase { -039 private boolean foundKV = false; -040 -041 public FirstKeyOnlyFilter() { -042 } -043 -044 @Override -045 public void reset() { -046foundKV = false; -047 } -048 -049 @Override -050 public boolean filterRowKey(Cell cell) throws IOException { -051// Impl in FilterBase might do unnecessary copy for Off heap backed Cells. -052return false; -053 } -054 -055 @Deprecated -056 @Override -057 public ReturnCode filterKeyValue(final Cell c) { -058return filterCell(c); -059 } -060 -061 @Override -062 public ReturnCode filterCell(final Cell c) { -063if(foundKV) return ReturnCode.NEXT_ROW; -064foundKV = true; -065return ReturnCode.INCLUDE; -066 } -067 -068 public static Filter createFilterFromArguments(ArrayListbyte [] filterArguments) { -069 Preconditions.checkArgument(filterArguments.isEmpty(), -070"Expected 0 but got: %s", filterArguments.size()); -071return new FirstKeyOnlyFilter(); -072 } -073 -074 /** -075 * @return true if first KV has been found. -076 */ -077 protected boolean hasFoundKV() { -078return this.foundKV; -079 } -080 -081 /** -082 * -083 * @param value update {@link #foundKV} flag with value. -084 */ -085 protected void setFoundKV(boolean value) { -086this.foundKV = value; -087 } -088 -089 /** -090 * @return The filter serialized using pb -091 */ -092 @Override -093 public byte [] toByteArray() { -094 FilterProtos.FirstKeyOnlyFilter.Builder builder = -095 FilterProtos.FirstKeyOnlyFilter.newBuilder(); -096return builder.build().toByteArray(); -097 } -098 -099 /** -100 * @param pbBytes A pb serialized {@link FirstKeyOnlyFilter} instance -101 * @return An instance of {@link FirstKeyOnlyFilter} made from codebytes/code -102 * @throws org.apache.hadoop.hbase.exceptions.DeserializationException -103 * @see #toByteArray -104 */ -105 public static FirstKeyOnlyFilter parseFrom(final byte [] pbBytes) -106 throws DeserializationException { -107// There is nothing to deserialize. Why do this at all? -108try { -109 FilterProtos.FirstKeyOnlyFilter.parseFrom(pbBytes); -110} catch (InvalidProtocolBufferException e) { -111 throw new DeserializationException(e); -112} -113// Just return a new instance. -114return new FirstKeyOnlyFilter(); -115 } -116 -117 /** -118 * @param o the other filter to compare with -119 * @return true if and only if the fields of the filter that are serialized -120 * are equal to the corresponding fields in other. Used for testing. -121 */ -122 @Override -123 boolean areSerializedFieldsEqual(Filter o) { -124if (o == this) return true; -125if (!(o instanceof FirstKeyOnlyFilter)) return false; -126 -127return true; -128 } -129} +023import java.util.Objects; +024 +025import org.apache.hadoop.hbase.Cell; +026import org.apache.yetus.audience.InterfaceAudience; +027import org.apache.hadoop.hbase.exceptions.DeserializationException; +028import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos; +029 +030import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; +031import org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException; +032 +033/** +034 * A filter that will only return the first KV from each row. +035 * p +036 * This filter can be used to more efficiently perform row count operations. +037 */ +038@InterfaceAudience.Public +039public class FirstKeyOnlyFilter extends FilterBase { +040 private boolean foundKV = false; +041 +042 public FirstKeyOnlyFilter() { +043 } +044 +045
[27/51] [partial] hbase-site git commit: Published site at 7c1fad4992a169a35b4457e6f4afcb30d04406e9.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/74f60271/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html b/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html index 71b9375..8f7d034 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html @@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab"; -private final class HStore.StoreFlusherImpl +private final class HStore.StoreFlusherImpl extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements StoreFlushContext @@ -279,7 +279,7 @@ implements tracker -private finalFlushLifeCycleTracker tracker +private finalFlushLifeCycleTracker tracker @@ -288,7 +288,7 @@ implements cacheFlushSeqNum -private finallong cacheFlushSeqNum +private finallong cacheFlushSeqNum @@ -297,7 +297,7 @@ implements snapshot -privateMemStoreSnapshot snapshot +privateMemStoreSnapshot snapshot @@ -306,7 +306,7 @@ implements tempFiles -privatehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.fs.Path tempFiles +privatehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.fs.Path tempFiles @@ -315,7 +315,7 @@ implements committedFiles -privatehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.fs.Path committedFiles +privatehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.fs.Path committedFiles @@ -324,7 +324,7 @@ implements cacheFlushCount -privatelong cacheFlushCount +privatelong cacheFlushCount @@ -333,7 +333,7 @@ implements cacheFlushSize -privatelong cacheFlushSize +privatelong cacheFlushSize @@ -342,7 +342,7 @@ implements outputFileSize -privatelong outputFileSize +privatelong outputFileSize @@ -359,7 +359,7 @@ implements StoreFlusherImpl -privateStoreFlusherImpl(longcacheFlushSeqNum, +privateStoreFlusherImpl(longcacheFlushSeqNum, FlushLifeCycleTrackertracker) @@ -377,7 +377,7 @@ implements prepare -publicMemStoreSizeprepare() +publicMemStoreSizeprepare() This is not thread safe. The caller should have a lock on the region or the store. If necessary, the lock can be added with the patch provided in HBASE-10087 @@ -394,7 +394,7 @@ implements flushCache -publicvoidflushCache(MonitoredTaskstatus) +publicvoidflushCache(MonitoredTaskstatus) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Description copied from interface:StoreFlushContext Flush the cache (create the new store file) @@ -415,7 +415,7 @@ implements commit -publicbooleancommit(MonitoredTaskstatus) +publicbooleancommit(MonitoredTaskstatus) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Description copied from interface:StoreFlushContext Commit the flush - add the store file to the store and clear the @@ -440,7 +440,7 @@ implements getOutputFileSize -publiclonggetOutputFileSize() +publiclonggetOutputFileSize() Specified by: getOutputFileSizein interfaceStoreFlushContext @@ -455,7 +455,7 @@ implements getCommittedFiles -publichttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.fs.PathgetCommittedFiles() +publichttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.fs.PathgetCommittedFiles() Description copied from interface:StoreFlushContext Returns the newly committed files from the flush. Called only if commit returns true @@ -472,7 +472,7 @@ implements replayFlush -publicvoidreplayFlush(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringfileNames, +publicvoidreplayFlush(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
[27/51] [partial] hbase-site git commit: Published site at 3afe9fb7e6ebfa71187cbe131558a83fae61cecd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html index e42e654..fe4e081 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html @@ -560,7 +560,7 @@ 552TransitRegionStateProcedure proc; 553regionNode.lock(); 554try { -555 preTransitCheck(regionNode, RegionStates.STATES_EXPECTED_ON_OPEN); +555 preTransitCheck(regionNode, STATES_EXPECTED_ON_ASSIGN); 556 proc = TransitRegionStateProcedure.assign(getProcedureEnvironment(), regionInfo, sn); 557 regionNode.setProcedure(proc); 558} finally { @@ -581,7 +581,7 @@ 573TransitRegionStateProcedure proc; 574regionNode.lock(); 575try { -576 preTransitCheck(regionNode, RegionStates.STATES_EXPECTED_ON_CLOSE); +576 preTransitCheck(regionNode, STATES_EXPECTED_ON_UNASSIGN_OR_MOVE); 577 proc = TransitRegionStateProcedure.unassign(getProcedureEnvironment(), regionInfo); 578 regionNode.setProcedure(proc); 579} finally { @@ -599,7 +599,7 @@ 591TransitRegionStateProcedure proc; 592regionNode.lock(); 593try { -594 preTransitCheck(regionNode, RegionStates.STATES_EXPECTED_ON_CLOSE); +594 preTransitCheck(regionNode, STATES_EXPECTED_ON_UNASSIGN_OR_MOVE); 595 regionNode.checkOnline(); 596 proc = TransitRegionStateProcedure.move(getProcedureEnvironment(), regionInfo, targetServer); 597 regionNode.setProcedure(proc); @@ -1419,448 +1419,483 @@ 1411 } 1412 1413 // -1414 // Region Status update -1415 // Should only be called in TransitRegionStateProcedure -1416 // -1417 private void transitStateAndUpdate(RegionStateNode regionNode, RegionState.State newState, -1418 RegionState.State... expectedStates) throws IOException { -1419RegionState.State state = regionNode.getState(); -1420regionNode.transitionState(newState, expectedStates); -1421boolean succ = false; -1422try { -1423 regionStateStore.updateRegionLocation(regionNode); -1424 succ = true; -1425} finally { -1426 if (!succ) { -1427// revert -1428regionNode.setState(state); -1429 } -1430} -1431 } -1432 -1433 // should be called within the synchronized block of RegionStateNode -1434 void regionOpening(RegionStateNode regionNode) throws IOException { -1435transitStateAndUpdate(regionNode, State.OPENING, RegionStates.STATES_EXPECTED_ON_OPEN); -1436 regionStates.addRegionToServer(regionNode); -1437// update the operation count metrics -1438 metrics.incrementOperationCounter(); -1439 } -1440 -1441 // should be called within the synchronized block of RegionStateNode. -1442 // The parameter 'giveUp' means whether we will try to open the region again, if it is true, then -1443 // we will persist the FAILED_OPEN state into hbase:meta. -1444 void regionFailedOpen(RegionStateNode regionNode, boolean giveUp) throws IOException { -1445RegionState.State state = regionNode.getState(); -1446ServerName regionLocation = regionNode.getRegionLocation(); -1447if (giveUp) { -1448 regionNode.setState(State.FAILED_OPEN); -1449 regionNode.setRegionLocation(null); -1450 boolean succ = false; -1451 try { -1452 regionStateStore.updateRegionLocation(regionNode); -1453succ = true; -1454 } finally { -1455if (!succ) { -1456 // revert -1457 regionNode.setState(state); -1458 regionNode.setRegionLocation(regionLocation); -1459} -1460 } -1461} -1462if (regionLocation != null) { -1463 regionStates.removeRegionFromServer(regionLocation, regionNode); -1464} -1465 } -1466 -1467 // should be called within the synchronized block of RegionStateNode -1468 void regionOpened(RegionStateNode regionNode) throws IOException { -1469// TODO: OPENING Updates hbase:meta too... we need to do both here and there? -1470// That is a lot of hbase:meta writing. -1471transitStateAndUpdate(regionNode, State.OPEN, RegionStates.STATES_EXPECTED_ON_OPEN); -1472RegionInfo hri = regionNode.getRegionInfo(); -1473if (isMetaRegion(hri)) { -1474 // Usually we'd set a table ENABLED at this stage but hbase:meta is ALWAYs enabled, it -1475 // can't be disabled -- so skip the RPC (besides... enabled is managed by TableStateManager
[27/51] [partial] hbase-site git commit: Published site at a452487a9b82bfd33bc10683c3f8b8ae74d58883.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cf79db0/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html b/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html index 2db4997..14fcebb 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html @@ -29,146 +29,161 @@ 021 022import java.io.IOException; 023import java.util.ArrayList; -024 -025import org.apache.hadoop.hbase.Cell; -026import org.apache.hadoop.hbase.CompareOperator; -027import org.apache.yetus.audience.InterfaceAudience; -028import org.apache.hadoop.hbase.exceptions.DeserializationException; -029import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -030import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos; -031import org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException; -032 -033/** -034 * This filter is used to filter based on the key. It takes an operator -035 * (equal, greater, not equal, etc) and a byte [] comparator for the row, -036 * and column qualifier portions of a key. -037 * p -038 * This filter can be wrapped with {@link WhileMatchFilter} to add more control. -039 * p -040 * Multiple filters can be combined using {@link FilterList}. -041 * p -042 * If an already known row range needs to be scanned, -043 * use {@link org.apache.hadoop.hbase.CellScanner} start -044 * and stop rows directly rather than a filter. -045 */ -046@InterfaceAudience.Public -047public class RowFilter extends CompareFilter { -048 -049 private boolean filterOutRow = false; -050 -051 /** -052 * Constructor. -053 * @param rowCompareOp the compare op for row matching -054 * @param rowComparator the comparator for row matching -055 * @deprecated Since 2.0.0. Will remove in 3.0.0. Use -056 * {@link #RowFilter(CompareOperator, ByteArrayComparable)}} instead. -057 */ -058 @Deprecated -059 public RowFilter(final CompareOp rowCompareOp, -060 final ByteArrayComparable rowComparator) { -061super(rowCompareOp, rowComparator); -062 } -063 -064 /** -065 * Constructor. -066 * @param op the compare op for row matching -067 * @param rowComparator the comparator for row matching -068 */ -069 public RowFilter(final CompareOperator op, -070 final ByteArrayComparable rowComparator) { -071super(op, rowComparator); -072 } -073 -074 @Override -075 public void reset() { -076this.filterOutRow = false; -077 } -078 -079 @Deprecated -080 @Override -081 public ReturnCode filterKeyValue(final Cell c) { -082return filterCell(c); -083 } -084 -085 @Override -086 public ReturnCode filterCell(final Cell v) { -087if(this.filterOutRow) { -088 return ReturnCode.NEXT_ROW; -089} -090return ReturnCode.INCLUDE; -091 } -092 -093 @Override -094 public boolean filterRowKey(Cell firstRowCell) { -095if (compareRow(getCompareOperator(), this.comparator, firstRowCell)) { -096 this.filterOutRow = true; -097} -098return this.filterOutRow; -099 } -100 -101 @Override -102 public boolean filterRow() { -103return this.filterOutRow; -104 } -105 -106 public static Filter createFilterFromArguments(ArrayListbyte [] filterArguments) { -107@SuppressWarnings("rawtypes") // for arguments -108ArrayList arguments = CompareFilter.extractArguments(filterArguments); -109CompareOperator compareOp = (CompareOperator)arguments.get(0); -110ByteArrayComparable comparator = (ByteArrayComparable)arguments.get(1); -111return new RowFilter(compareOp, comparator); -112 } -113 -114 /** -115 * @return The filter serialized using pb -116 */ -117 @Override -118 public byte [] toByteArray() { -119FilterProtos.RowFilter.Builder builder = -120 FilterProtos.RowFilter.newBuilder(); -121 builder.setCompareFilter(super.convert()); -122return builder.build().toByteArray(); -123 } -124 -125 /** -126 * @param pbBytes A pb serialized {@link RowFilter} instance -127 * @return An instance of {@link RowFilter} made from codebytes/code -128 * @throws DeserializationException -129 * @see #toByteArray -130 */ -131 public static RowFilter parseFrom(final byte [] pbBytes) -132 throws DeserializationException { -133FilterProtos.RowFilter proto; -134try { -135 proto = FilterProtos.RowFilter.parseFrom(pbBytes); -136} catch (InvalidProtocolBufferException e) { -137 throw new DeserializationException(e); -138} -139final CompareOperator valueCompareOp = -140 CompareOperator.valueOf(proto.getCompareFilter().getCompareOp().name()); -141ByteArrayComparable valueComparator = null; -142try { -143 if (proto.getCompareFilter().hasComparator()) { -144valueComparator =
[27/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html index db8431b..a8cb7c4 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html @@ -885,7766 +885,7797 @@ 877 * @return What the next sequence (edit) id should be. 878 * @throws IOException e 879 */ -880 private long initialize(final CancelableProgressable reporter) throws IOException { -881 -882//Refuse to open the region if there is no column family in the table -883if (htableDescriptor.getColumnFamilyCount() == 0) { -884 throw new DoNotRetryIOException("Table " + htableDescriptor.getTableName().getNameAsString()+ -885 " should have at least one column family."); -886} -887 -888MonitoredTask status = TaskMonitor.get().createStatus("Initializing region " + this); -889long nextSeqId = -1; -890try { -891 nextSeqId = initializeRegionInternals(reporter, status); -892 return nextSeqId; -893} finally { -894 // nextSeqid will be -1 if the initialization fails. -895 // At least it will be 0 otherwise. -896 if (nextSeqId == -1) { -897status.abort("Exception during region " + getRegionInfo().getRegionNameAsString() + -898 " initialization."); -899 } -900} -901 } -902 -903 private long initializeRegionInternals(final CancelableProgressable reporter, -904 final MonitoredTask status) throws IOException { -905if (coprocessorHost != null) { -906 status.setStatus("Running coprocessor pre-open hook"); -907 coprocessorHost.preOpen(); -908} -909 -910// Write HRI to a file in case we need to recover hbase:meta -911// Only the primary replica should write .regioninfo -912if (this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) { -913 status.setStatus("Writing region info on filesystem"); -914 fs.checkRegionInfoOnFilesystem(); -915} -916 -917// Initialize all the HStores -918status.setStatus("Initializing all the Stores"); -919long maxSeqId = initializeStores(reporter, status); -920this.mvcc.advanceTo(maxSeqId); -921if (ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) { -922 CollectionHStore stores = this.stores.values(); -923 try { -924// update the stores that we are replaying -925LOG.debug("replaying wal for " + this.getRegionInfo().getEncodedName()); -926 stores.forEach(HStore::startReplayingFromWAL); -927// Recover any edits if available. -928maxSeqId = Math.max(maxSeqId, -929 replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, status)); -930// Make sure mvcc is up to max. -931this.mvcc.advanceTo(maxSeqId); -932 } finally { -933LOG.debug("stopping wal replay for " + this.getRegionInfo().getEncodedName()); -934// update the stores that we are done replaying -935 stores.forEach(HStore::stopReplayingFromWAL); -936 } -937} -938this.lastReplayedOpenRegionSeqId = maxSeqId; -939 -940 this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this)); -941this.writestate.flushRequested = false; -942this.writestate.compacting.set(0); -943 -944if (this.writestate.writesEnabled) { -945 LOG.debug("Cleaning up temporary data for " + this.getRegionInfo().getEncodedName()); -946 // Remove temporary data left over from old regions -947 status.setStatus("Cleaning up temporary data from old regions"); -948 fs.cleanupTempDir(); -949} -950 -951if (this.writestate.writesEnabled) { -952 status.setStatus("Cleaning up detritus from prior splits"); -953 // Get rid of any splits or merges that were lost in-progress. Clean out -954 // these directories here on open. We may be opening a region that was -955 // being split but we crashed in the middle of it all. -956 LOG.debug("Cleaning up detritus for " + this.getRegionInfo().getEncodedName()); -957 fs.cleanupAnySplitDetritus(); -958 fs.cleanupMergesDir(); -959} +880 @VisibleForTesting +881 long initialize(final CancelableProgressable reporter) throws IOException { +882 +883//Refuse to open the region if there is no column family in the table +884if (htableDescriptor.getColumnFamilyCount() == 0) { +885 throw new DoNotRetryIOException("Table " + htableDescriptor.getTableName().getNameAsString()+ +886 " should have at least one column family."); +887}
[27/51] [partial] hbase-site git commit: Published site at 63f2d3cbdc8151f5f61f33e0a078c51b9ac076a5.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ae6a80c/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html index 8ffeb95..b1fbd27 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionTransitionProcedure.html @@ -18,8 +18,8 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":6,"i5":10,"i6":10,"i7":10,"i8":10,"i9":6,"i10":10,"i11":10,"i12":10,"i13":10,"i14":6,"i15":10,"i16":10,"i17":10,"i18":6,"i19":6,"i20":10,"i21":10,"i22":10,"i23":6,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":6,"i32":10,"i33":10,"i34":6,"i35":10}; -var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],8:["t4","Concrete Methods"]}; +var methods = {"i0":42,"i1":42,"i2":38,"i3":42,"i4":42,"i5":42,"i6":42,"i7":42,"i8":42,"i9":38,"i10":38,"i11":38,"i12":42,"i13":42,"i14":42,"i15":38,"i16":42,"i17":42,"i18":42,"i19":42,"i20":38,"i21":42,"i22":42,"i23":38}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],8:["t4","Concrete Methods"],32:["t6","Deprecated Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; var tableTab = "tableTab"; @@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab"; PrevClass -NextClass +NextClass Frames @@ -121,58 +121,20 @@ var activeTableTab = "activeTableTab"; AssignProcedure, UnassignProcedure +Deprecated. +Do not use any more. + -@InterfaceAudience.Private -public abstract class RegionTransitionProcedure +https://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true; title="class or interface in java.lang">@Deprecated + @InterfaceAudience.Private +public abstract class RegionTransitionProcedure extends ProcedureMasterProcedureEnv implements TableProcedureInterface, RemoteProcedureDispatcher.RemoteProcedureMasterProcedureEnv,ServerName -Base class for the Assign and Unassign Procedure. - - Locking: - Takes exclusive lock on the region being assigned/unassigned. Thus, there can only be one - RegionTransitionProcedure per region running at a time (see MasterProcedureScheduler). - - This procedure is asynchronous and responds to external events. - The AssignmentManager will notify this procedure when the RS completes - the operation and reports the transitioned state - (see the Assign and Unassign class for more detail). - - Procedures move from the REGION_TRANSITION_QUEUE state when they are - first submitted, to the REGION_TRANSITION_DISPATCH state when the request - to remote server is sent and the Procedure is suspended waiting on external - event to be woken again. Once the external event is triggered, Procedure - moves to the REGION_TRANSITION_FINISH state. - - NOTE: AssignProcedure and UnassignProcedure should not be thought of - as being asymmetric, at least currently. - - AssignProcedure moves through all the above described states and implements methods - associated with each while UnassignProcedure starts at state - REGION_TRANSITION_DISPATCH and state REGION_TRANSITION_QUEUE is not supported. - - When any step in AssignProcedure fails, failure handler - AssignProcedure#handleFailure(MasterProcedureEnv, RegionStateNode) re-attempts the - assignment by setting the procedure state to REGION_TRANSITION_QUEUE and forces - assignment to a different target server by setting AssignProcedure.forceNewPlan. When - the number of attempts reaches threshold configuration 'hbase.assignment.maximum.attempts', - the procedure is aborted. For UnassignProcedure, similar re-attempts are - intentionally not implemented. It is a 'one shot' procedure. See its class doc for how it - handles failure. - - If we find a region in an 'unexpected' state, we'll complain and retry with backoff forever. - The 'unexpected' state needs to be fixed either by another running Procedure or by operator - intervention (Regions in 'unexpected' state indicates bug or unexpected transition type). - For this to work, subclasses need to persist the 'attempt' counter kept in this class when - they do serializeStateData and restore it inside their deserializeStateData, just as they do - for regionInfo. - - - - - TODO: Considering it is a priority doing all we can to get make a region available as soon as - possible, re-attempting with any target makes sense if specified target fails in case of - AssignProcedure. For UnassignProcedure, our concern is preventing data loss - on failed unassign. See class doc for explanation. +Leave here only for checking if we can successfully start the master. + +See Also:
[27/51] [partial] hbase-site git commit: Published site at 092efb42749bf7fc6ad338c96aae8e7b9d3a2c74.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f3d62514/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellWritableComparablePartitioner.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellWritableComparablePartitioner.html b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellWritableComparablePartitioner.html index 39170f0..7859ebc 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellWritableComparablePartitioner.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.CellWritableComparablePartitioner.html @@ -230,564 +230,567 @@ 222 } 223} 224 } catch (InterruptedException e) { -225e.printStackTrace(); -226 } -227} -228 -229@Override -230public void setup(Context context) throws IOException { -231 cfRenameMap = createCfRenameMap(context.getConfiguration()); -232 filter = instantiateFilter(context.getConfiguration()); -233 int reduceNum = context.getNumReduceTasks(); -234 Configuration conf = context.getConfiguration(); -235 TableName tableName = TableName.valueOf(context.getConfiguration().get(TABLE_NAME)); -236 try (Connection conn = ConnectionFactory.createConnection(conf); -237 RegionLocator regionLocator = conn.getRegionLocator(tableName)) { -238byte[][] startKeys = regionLocator.getStartKeys(); -239if (startKeys.length != reduceNum) { -240 throw new IOException("Region split after job initialization"); -241} -242CellWritableComparable[] startKeyWraps = -243new CellWritableComparable[startKeys.length - 1]; -244for (int i = 1; i startKeys.length; ++i) { -245 startKeyWraps[i - 1] = -246 new CellWritableComparable(KeyValueUtil.createFirstOnRow(startKeys[i])); -247} -248 CellWritableComparablePartitioner.START_KEYS = startKeyWraps; -249 } -250} -251 } -252 -253 /** -254 * A mapper that just writes out KeyValues. -255 */ -256 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="EQ_COMPARETO_USE_OBJECT_EQUALS", -257 justification="Writables are going away and this has been this way forever") -258 public static class CellImporter extends TableMapperImmutableBytesWritable, Cell { -259private Mapbyte[], byte[] cfRenameMap; -260private Filter filter; -261private static final Logger LOG = LoggerFactory.getLogger(CellImporter.class); -262 -263/** -264 * @param row The current table row key. -265 * @param value The columns. -266 * @param context The current context. -267 * @throws IOException When something is broken with the data. -268 */ -269@Override -270public void map(ImmutableBytesWritable row, Result value, -271 Context context) -272throws IOException { -273 try { -274if (LOG.isTraceEnabled()) { -275 LOG.trace("Considering the row." -276 + Bytes.toString(row.get(), row.getOffset(), row.getLength())); -277} -278if (filter == null -279|| !filter.filterRowKey(PrivateCellUtil.createFirstOnRow(row.get(), row.getOffset(), -280(short) row.getLength( { -281 for (Cell kv : value.rawCells()) { -282kv = filterKv(filter, kv); -283// skip if we filtered it out -284if (kv == null) continue; -285context.write(row, new MapReduceExtendedCell(convertKv(kv, cfRenameMap))); -286 } -287} -288 } catch (InterruptedException e) { -289e.printStackTrace(); -290 } -291} -292 -293@Override -294public void setup(Context context) { -295 cfRenameMap = createCfRenameMap(context.getConfiguration()); -296 filter = instantiateFilter(context.getConfiguration()); -297} -298 } -299 -300 /** -301 * Write table content out to files in hdfs. -302 */ -303 public static class Importer extends TableMapperImmutableBytesWritable, Mutation { -304private Mapbyte[], byte[] cfRenameMap; -305private ListUUID clusterIds; -306private Filter filter; -307private Durability durability; -308 -309/** -310 * @param row The current table row key. -311 * @param value The columns. -312 * @param context The current context. -313 * @throws IOException When something is broken with the data. -314 */ -315@Override -316public void map(ImmutableBytesWritable row, Result value, -317 Context context) -318throws IOException { -319 try { -320writeResult(row, value, context); -321 } catch (InterruptedException e) { -322e.printStackTrace(); -323 } -324} -325 -326private void writeResult(ImmutableBytesWritable key, Result result, Context context) -327throws IOException,
[27/51] [partial] hbase-site git commit: Published site at 613d831429960348dc42c3bdb6ea5d31be15c81c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReader.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReader.html b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReader.html index b7b4236..3d1edb3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReader.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.FSReader.html @@ -259,1863 +259,1867 @@ 251 * + Metadata! + = See note on BLOCK_METADATA_SPACE above. 252 * ++ 253 * /code -254 * @see #serialize(ByteBuffer) +254 * @see #serialize(ByteBuffer, boolean) 255 */ -256 static final CacheableDeserializerCacheable BLOCK_DESERIALIZER = -257 new CacheableDeserializerCacheable() { -258@Override -259public HFileBlock deserialize(ByteBuff buf, boolean reuse, MemoryType memType) -260throws IOException { -261 // The buf has the file block followed by block metadata. -262 // Set limit to just before the BLOCK_METADATA_SPACE then rewind. -263 buf.limit(buf.limit() - BLOCK_METADATA_SPACE).rewind(); -264 // Get a new buffer to pass the HFileBlock for it to 'own'. -265 ByteBuff newByteBuff; -266 if (reuse) { -267newByteBuff = buf.slice(); -268 } else { -269int len = buf.limit(); -270newByteBuff = new SingleByteBuff(ByteBuffer.allocate(len)); -271newByteBuff.put(0, buf, buf.position(), len); -272 } -273 // Read out the BLOCK_METADATA_SPACE content and shove into our HFileBlock. -274 buf.position(buf.limit()); -275 buf.limit(buf.limit() + HFileBlock.BLOCK_METADATA_SPACE); -276 boolean usesChecksum = buf.get() == (byte) 1; -277 long offset = buf.getLong(); -278 int nextBlockOnDiskSize = buf.getInt(); -279 HFileBlock hFileBlock = -280 new HFileBlock(newByteBuff, usesChecksum, memType, offset, nextBlockOnDiskSize, null); -281 return hFileBlock; -282} -283 -284@Override -285public int getDeserialiserIdentifier() { -286 return DESERIALIZER_IDENTIFIER; -287} -288 -289@Override -290public HFileBlock deserialize(ByteBuff b) throws IOException { -291 // Used only in tests -292 return deserialize(b, false, MemoryType.EXCLUSIVE); -293} -294 }; -295 -296 private static final int DESERIALIZER_IDENTIFIER; -297 static { -298DESERIALIZER_IDENTIFIER = -299 CacheableDeserializerIdManager.registerDeserializer(BLOCK_DESERIALIZER); -300 } -301 -302 /** -303 * Copy constructor. Creates a shallow copy of {@code that}'s buffer. -304 */ -305 private HFileBlock(HFileBlock that) { -306this(that, false); -307 } -308 -309 /** -310 * Copy constructor. Creates a shallow/deep copy of {@code that}'s buffer as per the boolean -311 * param. -312 */ -313 private HFileBlock(HFileBlock that, boolean bufCopy) { -314init(that.blockType, that.onDiskSizeWithoutHeader, -315 that.uncompressedSizeWithoutHeader, that.prevBlockOffset, -316that.offset, that.onDiskDataSizeWithHeader, that.nextBlockOnDiskSize, that.fileContext); -317if (bufCopy) { -318 this.buf = new SingleByteBuff(ByteBuffer.wrap(that.buf.toBytes(0, that.buf.limit(; -319} else { -320 this.buf = that.buf.duplicate(); -321} -322 } -323 -324 /** -325 * Creates a new {@link HFile} block from the given fields. This constructor -326 * is used only while writing blocks and caching, -327 * and is sitting in a byte buffer and we want to stuff the block into cache. -328 * -329 * pTODO: The caller presumes no checksumming -330 * required of this block instance since going into cache; checksum already verified on -331 * underlying block data pulled in from filesystem. Is that correct? What if cache is SSD? +256 public static final CacheableDeserializerCacheable BLOCK_DESERIALIZER = new BlockDeserializer(); +257 +258 public static final class BlockDeserializer implements CacheableDeserializerCacheable { +259private BlockDeserializer() { +260} +261 +262@Override +263public HFileBlock deserialize(ByteBuff buf, boolean reuse, MemoryType memType) +264throws IOException { +265 // The buf has the file block followed by block metadata. +266 // Set limit to just before the BLOCK_METADATA_SPACE then rewind. +267 buf.limit(buf.limit() - BLOCK_METADATA_SPACE).rewind(); +268 // Get a new buffer to pass the HFileBlock for it to 'own'. +269 ByteBuff newByteBuff; +270 if (reuse) { +271newByteBuff = buf.slice(); +272 } else { +273int len = buf.limit(); +274newByteBuff = new SingleByteBuff(ByteBuffer.allocate(len)); +275newByteBuff.put(0, buf, buf.position(), len); +276 } +277
[27/51] [partial] hbase-site git commit: Published site at ba5d1c1f28301adc99019d9d6c4a04fac98ae511.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/804782f0/devapidocs/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.html index 9d1748f..ac41873 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/RestoreSnapshotProcedure.html @@ -411,7 +411,7 @@ extends AbstractStateMachineTableProcedure -acquireLock, checkOnline, checkTableModifiable, getRegionDir, getUser, preflightChecks, releaseLock, releaseSyncLatch, setUser +acquireLock, checkOnline, checkTableModifiable, getRegionDir, getUser, preflightChecks, releaseLock, releaseSyncLatch, setUser, waitInitialized< /li> @@ -425,7 +425,7 @@ extends Procedure -addStackIndex, afterReplay, beforeReplay, compareTo, completionCleanup, doAcquireLock, doExecute, d oReleaseLock, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isFailed, isFinished, isInitializing, isRunnable, isSuccess, < a href="../../../../../../org/apache/hadoop/hbase/procedure2/Procedure.html#isWaiting--">isWaiting, removeStackIndex, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, shouldWaitClientAck, toString, toStringClass, toStringDetails, toStringSimpleSB, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, wasExecuted +addStackIndex, afterReplay, beforeReplay, compareTo, completionCleanup, doExecute, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, removeStackIndex, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, shouldWaitClientAck, toString, toStringClass, toStringDetails, toStringSimpleSB, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, wasExecuted http://git-wip-us.apache.org/repos/asf/hbase-site/blob/804782f0/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.html index ad95213..941b155 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashProcedure.html @@ -411,7 +411,7 @@ implements Procedure -addStackIndex, afterReplay, beforeReplay, compareTo, completionCleanup, doAcquireLock, doExecute, d oReleaseLock, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcId, getProcIdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, removeStackIndex, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, toString, toStringClass, toStringDetails, toStringSimpleSB, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, wasExecuted +addStackIndex, afterReplay, beforeReplay, compareTo, completionCleanup, doExecute, doRollback,
[27/51] [partial] hbase-site git commit: Published site at b4759ce6e72f50ccd9d410bd5917dc5a515414f1.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/df8fd1d3/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadAsRateFunction.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadAsRateFunction.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadAsRateFunction.html index 233dba3..91b9055 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadAsRateFunction.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CostFromRegionLoadAsRateFunction.html @@ -540,1205 +540,1204 @@ 532 sm.getRegionMetrics().forEach((byte[] regionName, RegionMetrics rm) - { 533DequeBalancerRegionLoad rLoads = oldLoads.get(Bytes.toString(regionName)); 534if (rLoads == null) { -535 // There was nothing there -536 rLoads = new ArrayDeque(); -537} else if (rLoads.size() = numRegionLoadsToRemember) { -538 rLoads.remove(); -539} -540rLoads.add(new BalancerRegionLoad(rm)); -541 loads.put(Bytes.toString(regionName), rLoads); -542 }); -543}); -544 -545for(CostFromRegionLoadFunction cost : regionLoadFunctions) { -546 cost.setLoads(loads); -547} -548 } -549 -550 protected void initCosts(Cluster cluster) { -551for (CostFunction c:costFunctions) { -552 c.init(cluster); -553} -554 } -555 -556 protected void updateCostsWithAction(Cluster cluster, Action action) { -557for (CostFunction c : costFunctions) { -558 c.postAction(action); -559} -560 } -561 -562 /** -563 * Get the names of the cost functions -564 */ -565 public String[] getCostFunctionNames() { -566if (costFunctions == null) return null; -567String[] ret = new String[costFunctions.length]; -568for (int i = 0; i costFunctions.length; i++) { -569 CostFunction c = costFunctions[i]; -570 ret[i] = c.getClass().getSimpleName(); -571} -572 -573return ret; -574 } -575 -576 /** -577 * This is the main cost function. It will compute a cost associated with a proposed cluster -578 * state. All different costs will be combined with their multipliers to produce a double cost. -579 * -580 * @param cluster The state of the cluster -581 * @param previousCost the previous cost. This is used as an early out. -582 * @return a double of a cost associated with the proposed cluster state. This cost is an -583 * aggregate of all individual cost functions. -584 */ -585 protected double computeCost(Cluster cluster, double previousCost) { -586double total = 0; -587 -588for (int i = 0; i costFunctions.length; i++) { -589 CostFunction c = costFunctions[i]; -590 this.tempFunctionCosts[i] = 0.0; -591 -592 if (c.getMultiplier() = 0) { -593continue; -594 } -595 -596 Float multiplier = c.getMultiplier(); -597 Double cost = c.cost(); -598 -599 this.tempFunctionCosts[i] = multiplier*cost; -600 total += this.tempFunctionCosts[i]; -601 -602 if (total previousCost) { -603break; -604 } -605} -606 -607return total; -608 } -609 -610 /** Generates a candidate action to be applied to the cluster for cost function search */ -611 abstract static class CandidateGenerator { -612abstract Cluster.Action generate(Cluster cluster); -613 -614/** -615 * From a list of regions pick a random one. Null can be returned which -616 * {@link StochasticLoadBalancer#balanceCluster(Map)} recognize as signal to try a region move -617 * rather than swap. -618 * -619 * @param clusterThe state of the cluster -620 * @param server index of the server -621 * @param chanceOfNoSwap Chance that this will decide to try a move rather -622 * than a swap. -623 * @return a random {@link RegionInfo} or null if an asymmetrical move is -624 * suggested. -625 */ -626protected int pickRandomRegion(Cluster cluster, int server, double chanceOfNoSwap) { -627 // Check to see if this is just a move. -628 if (cluster.regionsPerServer[server].length == 0 || RANDOM.nextFloat() chanceOfNoSwap) { -629// signal a move only. -630return -1; -631 } -632 int rand = RANDOM.nextInt(cluster.regionsPerServer[server].length); -633 return cluster.regionsPerServer[server][rand]; -634 -635} -636protected int pickRandomServer(Cluster cluster) { -637 if (cluster.numServers 1) { -638return -1; -639 } -640 -641 return RANDOM.nextInt(cluster.numServers); -642} -643 -644protected int pickRandomRack(Cluster cluster) { -645 if
[27/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0c6f447e/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html -- diff --git a/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html b/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html index f1708c1..8a6ff31 100644 --- a/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html +++ b/apidocs/org/apache/hadoop/hbase/LocalHBaseCluster.html @@ -1,6 +1,6 @@ http://www.w3.org/TR/html4/loose.dtd;> - + @@ -19,45 +19,45 @@ } //--> var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":9,"i15":10,"i16":9,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10}; -var tabs = {65535:["t0","æææ¹æ³"],1:["t1","éææ¹æ³"],2:["t2","å®ä¾æ¹æ³"],8:["t4","å ·ä½æ¹æ³"]}; +var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; var tableTab = "tableTab"; var activeTableTab = "activeTableTab"; -æ¨çæµè§å¨å·²ç¦ç¨ JavaScriptã +JavaScript is disabled on your browser. -è·³è¿å¯¼èªé¾æ¥ +Skip navigation links - -æ¦è§ -ç¨åºå -ç±» -ä½¿ç¨ -æ -å·²è¿æ¶ -ç´¢å¼ -å¸®å© + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help -ä¸ä¸ä¸ªç±» -ä¸ä¸ä¸ªç±» +PrevClass +NextClass -æ¡æ¶ -æ æ¡æ¶ +Frames +NoFrames -ææç±» +AllClasses
[27/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5427a45e/apidocs/org/apache/hadoop/hbase/InvalidFamilyOperationException.html -- diff --git a/apidocs/org/apache/hadoop/hbase/InvalidFamilyOperationException.html b/apidocs/org/apache/hadoop/hbase/InvalidFamilyOperationException.html index c0db59c..74ecf35 100644 --- a/apidocs/org/apache/hadoop/hbase/InvalidFamilyOperationException.html +++ b/apidocs/org/apache/hadoop/hbase/InvalidFamilyOperationException.html @@ -1,6 +1,6 @@ http://www.w3.org/TR/html4/loose.dtd;> - + @@ -20,38 +20,38 @@ //--> -JavaScript is disabled on your browser. +æ¨çæµè§å¨å·²ç¦ç¨ JavaScriptã -Skip navigation links +è·³è¿å¯¼èªé¾æ¥ - -Overview -Package -Class -Use -Tree -Deprecated -Index -Help + +æ¦è§ +ç¨åºå +ç±» +ä½¿ç¨ +æ +å·²è¿æ¶ +ç´¢å¼ +å¸®å© -PrevClass -NextClass +ä¸ä¸ä¸ªç±» +ä¸ä¸ä¸ªç±» -Frames -NoFrames +æ¡æ¶ +æ æ¡æ¶ -AllClasses +ææç±»org.apache.hadoop.hbase-Class InvalidFamilyOperationException
+ç±» InvalidFamilyOperationException
-
- https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true" title="class or interface in java.lang">java.lang.Object
+- https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true" title="java.langä¸çç±»ææ¥å£">java.lang.Object
-
- https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true" title="class or interface in java.lang">java.lang.Throwable
+- https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true" title="java.langä¸çç±»ææ¥å£">java.lang.Throwable
-
- https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true" title="class or interface in java.lang">java.lang.Exception
+- https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true" title="java.langä¸çç±»ææ¥å£">java.lang.Exception
-
- https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">java.io.IOException
+- https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="java.ioä¸çç±»ææ¥å£">java.io.IOException
-
- org.apache.hadoop.hbase.HBaseIOException
+
[27/51] [partial] hbase-site git commit: Published site at 0f23784182ab88649de340d75804e0ff20dcd0fc.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bcb555af/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServicesForStores.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServicesForStores.html b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServicesForStores.html index e23e8f7..cf50fff 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServicesForStores.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServicesForStores.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -186,37 +186,29 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html longoffHeapSizeDelta) -void -blockUpdates() - - https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ThreadPoolExecutor.html?is-external=true; title="class or interface in java.util.concurrent">ThreadPoolExecutor getInMemoryCompactionPool() - + long getMemStoreFlushSize() - + (package private) long getMemStoreSize() - + int getNumStores() - + RegionInfo getRegionInfo() - + WAL getWAL() - -void -unblockUpdates() - @@ -295,31 +287,13 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html Method Detail - - - - - -blockUpdates -publicvoidblockUpdates() - - - - - - - -unblockUpdates -publicvoidunblockUpdates() - - addMemStoreSize -publicvoidaddMemStoreSize(longdataSizeDelta, +publicvoidaddMemStoreSize(longdataSizeDelta, longheapSizeDelta, longoffHeapSizeDelta) @@ -330,7 +304,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html getRegionInfo -publicRegionInfogetRegionInfo() +publicRegionInfogetRegionInfo() @@ -339,7 +313,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html getWAL -publicWALgetWAL() +publicWALgetWAL() @@ -348,7 +322,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html getInMemoryCompactionPool -publichttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ThreadPoolExecutor.html?is-external=true; title="class or interface in java.util.concurrent">ThreadPoolExecutorgetInMemoryCompactionPool() +publichttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ThreadPoolExecutor.html?is-external=true; title="class or interface in java.util.concurrent">ThreadPoolExecutorgetInMemoryCompactionPool() @@ -357,7 +331,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html getMemStoreFlushSize -publiclonggetMemStoreFlushSize() +publiclonggetMemStoreFlushSize() @@ -366,7 +340,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html getNumStores -publicintgetNumStores() +publicintgetNumStores() @@ -375,7 +349,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html getMemStoreSize -longgetMemStoreSize() +longgetMemStoreSize() http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bcb555af/devapidocs/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.html b/devapidocs/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.html index 84a1668..5d9b1ff 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":6,"i1":6,"i2":6}; +var methods = {"i0":6,"i1":6,"i2":6,"i3":6}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -110,7 +110,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public interface ReplicationSourceService +public interface ReplicationSourceService extends ReplicationService A source for a replication stream has to expose this service. This service allows an application to hook into the regionserver and watch for new transactions. @@ -139,12 +139,18 @@ extends +ReplicationSourceManager +getReplicationManager() +Returns the replication manager + + + ReplicationPeers getReplicationPeers() Return the replication peers. - + SyncReplicationPeerInfoProvider getSyncReplicationPeerInfoProvider() Returns an info provider for sync replication peer. @@ -178,7 +184,7 @@ extends
[27/51] [partial] hbase-site git commit: Published site at 85b41f36e01214b6485c9352875c84ebf877dab3.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a5c66de0/devapidocs/org/apache/hadoop/hbase/security/access/AccessControlUtil.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/security/access/AccessControlUtil.html b/devapidocs/org/apache/hadoop/hbase/security/access/AccessControlUtil.html index 2b98dff..a644a4a 100644 --- a/devapidocs/org/apache/hadoop/hbase/security/access/AccessControlUtil.html +++ b/devapidocs/org/apache/hadoop/hbase/security/access/AccessControlUtil.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9,"i7":9,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":9,"i26":9,"i27":9,"i28":9}; +var methods = {"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9,"i7":9,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":9,"i26":9,"i27":9,"i28":9,"i29":9,"i30":9,"i31":9,"i32":9,"i33":9}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -110,7 +110,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public class AccessControlUtil +public class AccessControlUtil extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object Since: @@ -229,13 +229,42 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListUserPermission +getUserPermissions(com.google.protobuf.RpcControllercontroller, + org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterfaceprotocol, + byte[]namespace, + https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringuserName) +A utility used to get permissions for selected namespace based on the specified user name. + + + +static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListUserPermission +getUserPermissions(com.google.protobuf.RpcControllercontroller, + org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterfaceprotocol, + https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringuserName) +A utility used to get user's global permissions based on the specified user name. + + + +static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListUserPermission getUserPermissions(com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterfaceprotocol, TableNamet) A utility used to get user table permissions. - + +static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListUserPermission +getUserPermissions(com.google.protobuf.RpcControllercontroller, + org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterfaceprotocol, + TableNamet, + byte[]columnFamily, + byte[]columnQualifier, + https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringuserName) +A utility used to get user table permissions based on the column family, column qualifier and + user name. + + + static void grant(com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterfaceprotocol, @@ -245,7 +274,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html A utility used to grant a user global permissions. - + static void grant(com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterfaceprotocol, @@ -256,7 +285,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html A utility used to grant a user namespace permissions. - + static void grant(com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterfaceprotocol, @@ -269,7 +298,20 @@ extends
[27/51] [partial] hbase-site git commit: Published site at 6198e1fc7dfa85c3bc6b2855f9a5fb5f4b2354ff.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb5d2c62/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html index 1410333..b56a22f 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html +++ b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationFuture.html @@ -124,7 +124,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private @InterfaceStability.Evolving -private static class HBaseAdmin.ReplicationFuture +private static class HBaseAdmin.ReplicationFuture extends HBaseAdmin.ProcedureFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void @@ -241,7 +241,7 @@ extends peerId -private finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String peerId +private finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String peerId @@ -250,7 +250,7 @@ extends getOperation -private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/function/Supplier.html?is-external=true; title="class or interface in java.util.function">Supplierhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String getOperation +private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/function/Supplier.html?is-external=true; title="class or interface in java.util.function">Supplierhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String getOperation @@ -267,7 +267,7 @@ extends ReplicationFuture -publicReplicationFuture(HBaseAdminadmin, +publicReplicationFuture(HBaseAdminadmin, https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringpeerId, https://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">LongprocId, https://docs.oracle.com/javase/8/docs/api/java/util/function/Supplier.html?is-external=true; title="class or interface in java.util.function">Supplierhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetOperation) @@ -287,7 +287,7 @@ extends toString -publichttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringtoString() +publichttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringtoString() Overrides: https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--; title="class or interface in java.lang">toStringin classhttps://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb5d2c62/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html index 9732695..356 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html +++ b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html @@ -127,7 +127,7 @@ var activeTableTab = "activeTableTab"; -private static class HBaseAdmin.RestoreSnapshotFuture +private static class HBaseAdmin.RestoreSnapshotFuture extends HBaseAdmin.TableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void @@ -240,7 +240,7 @@ extends RestoreSnapshotFuture -publicRestoreSnapshotFuture(HBaseAdminadmin, +publicRestoreSnapshotFuture(HBaseAdminadmin, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionsnapshot, TableNametableName, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponseresponse) @@ -252,7 +252,7 @@ extends RestoreSnapshotFuture -publicRestoreSnapshotFuture(HBaseAdminadmin, +publicRestoreSnapshotFuture(HBaseAdminadmin, TableNametableName,
[27/51] [partial] hbase-site git commit: Published site at 14087cc919da9f2e0b1a68f701f6365ad9d1d71f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/55ce8d97/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/MoveRegionProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/MoveRegionProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/MoveRegionProcedure.html index b0c0fe7..08be9d0 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/MoveRegionProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/MoveRegionProcedure.html @@ -86,6 +86,10 @@ org.apache.hadoop.hbase.master.assignment + +org.apache.hadoop.hbase.master.procedure + + @@ -103,12 +107,27 @@ -MoveRegionProcedure +private MoveRegionProcedure AssignmentManager.createMoveRegionProcedure(RegionPlanplan) - -MoveRegionProcedure[] -AssignmentManager.createReopenProcedures(https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true; title="class or interface in java.util">CollectionRegionInforegionInfo) + + + + + + +Uses of MoveRegionProcedure in org.apache.hadoop.hbase.master.procedure + +Methods in org.apache.hadoop.hbase.master.procedure that return MoveRegionProcedure + +Modifier and Type +Method and Description + + + +private MoveRegionProcedure +ReopenTableRegionsProcedure.createReopenProcedure(MasterProcedureEnvenv, + HRegionLocationloc) http://git-wip-us.apache.org/repos/asf/hbase-site/blob/55ce8d97/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html index f9b3413..a8b7714 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html @@ -245,40 +245,44 @@ RegionStates.RegionStateNode.compareTo(RegionStates.RegionStateNodeother) +private HRegionLocation +RegionStates.createRegionForReopen(RegionStates.RegionStateNodenode) + + protected void UnassignProcedure.finishTransition(MasterProcedureEnvenv, RegionStates.RegionStateNoderegionNode) - + protected void AssignProcedure.finishTransition(MasterProcedureEnvenv, RegionStates.RegionStateNoderegionNode) - + protected abstract void RegionTransitionProcedure.finishTransition(MasterProcedureEnvenv, RegionStates.RegionStateNoderegionNode) - + private void AssignProcedure.handleFailure(MasterProcedureEnvenv, RegionStates.RegionStateNoderegionNode) Called when dispatch or subsequent OPEN request fail. - + (package private) boolean RegionStates.include(RegionStates.RegionStateNodenode, booleanoffline) Utility. - + private boolean AssignProcedure.incrementAndCheckMaxAttempts(MasterProcedureEnvenv, RegionStates.RegionStateNoderegionNode) - + private boolean UnassignProcedure.isSafeToProceed(MasterProcedureEnvenv, RegionStates.RegionStateNoderegionNode, @@ -287,143 +291,143 @@ unassign; e.g. - + protected boolean RegionTransitionProcedure.isServerOnline(MasterProcedureEnvenv, RegionStates.RegionStateNoderegionNode) - + void AssignmentManager.markRegionAsClosed(RegionStates.RegionStateNoderegionNode) - + void AssignmentManager.markRegionAsClosing(RegionStates.RegionStateNoderegionNode) - + void AssignmentManager.markRegionAsOpened(RegionStates.RegionStateNoderegionNode) - + void AssignmentManager.markRegionAsOpening(RegionStates.RegionStateNoderegionNode) - + protected void UnassignProcedure.proceed(MasterProcedureEnvenv, RegionStates.RegionStateNoderegionNode) Set it up so when procedure is unsuspended, we'll move to the procedure finish. - + protected void AssignmentManager.queueAssign(RegionStates.RegionStateNoderegionNode) Add the assign operation to the assignment queue. - + protected boolean UnassignProcedure.remoteCallFailed(MasterProcedureEnvenv, RegionStates.RegionStateNoderegionNode, https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOExceptionexception) - + protected boolean AssignProcedure.remoteCallFailed(MasterProcedureEnvenv, RegionStates.RegionStateNoderegionNode, https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOExceptionexception) - + protected abstract boolean
[27/51] [partial] hbase-site git commit: Published site at 72784c2d836a4b977667449d3adec5e8d15453f5.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2b11656f/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DisableTableFuture.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DisableTableFuture.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DisableTableFuture.html index b6e7636..592c2cc 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DisableTableFuture.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DisableTableFuture.html @@ -356,3901 +356,3924 @@ 348 public FutureVoid modifyTableAsync(TableDescriptor td) throws IOException { 349ModifyTableResponse response = executeCallable( 350 new MasterCallableModifyTableResponse(getConnection(), getRpcControllerFactory()) { -351@Override -352protected ModifyTableResponse rpcCall() throws Exception { -353 setPriority(td.getTableName()); -354 ModifyTableRequest request = RequestConverter.buildModifyTableRequest( -355td.getTableName(), td, ng.getNonceGroup(), ng.newNonce()); -356 return master.modifyTable(getRpcController(), request); -357} -358 }); -359return new ModifyTableFuture(this, td.getTableName(), response); -360 } -361 -362 @Override -363 public ListTableDescriptor listTableDescriptorsByNamespace(byte[] name) throws IOException { -364return executeCallable(new MasterCallableListTableDescriptor(getConnection(), -365getRpcControllerFactory()) { -366 @Override -367 protected ListTableDescriptor rpcCall() throws Exception { -368return master.listTableDescriptorsByNamespace(getRpcController(), -369 ListTableDescriptorsByNamespaceRequest.newBuilder() -370 .setNamespaceName(Bytes.toString(name)).build()) -371.getTableSchemaList() -372.stream() -373 .map(ProtobufUtil::toTableDescriptor) -374 .collect(Collectors.toList()); -375 } -376}); -377 } -378 -379 @Override -380 public ListTableDescriptor listTableDescriptors(ListTableName tableNames) throws IOException { -381return executeCallable(new MasterCallableListTableDescriptor(getConnection(), -382getRpcControllerFactory()) { -383 @Override -384 protected ListTableDescriptor rpcCall() throws Exception { -385GetTableDescriptorsRequest req = -386 RequestConverter.buildGetTableDescriptorsRequest(tableNames); -387 return ProtobufUtil.toTableDescriptorList(master.getTableDescriptors(getRpcController(), -388 req)); -389 } -390}); -391 } -392 -393 @Override -394 public ListRegionInfo getRegions(final ServerName sn) throws IOException { -395AdminService.BlockingInterface admin = this.connection.getAdmin(sn); -396// TODO: There is no timeout on this controller. Set one! -397HBaseRpcController controller = rpcControllerFactory.newController(); -398return ProtobufUtil.getOnlineRegions(controller, admin); -399 } -400 -401 @Override -402 public ListRegionInfo getRegions(TableName tableName) throws IOException { -403if (TableName.isMetaTableName(tableName)) { -404 return Arrays.asList(RegionInfoBuilder.FIRST_META_REGIONINFO); -405} else { -406 return MetaTableAccessor.getTableRegions(connection, tableName, true); -407} -408 } -409 -410 private static class AbortProcedureFuture extends ProcedureFutureBoolean { -411private boolean isAbortInProgress; -412 -413public AbortProcedureFuture( -414final HBaseAdmin admin, -415final Long procId, -416final Boolean abortProcResponse) { -417 super(admin, procId); -418 this.isAbortInProgress = abortProcResponse; -419} -420 -421@Override -422public Boolean get(long timeout, TimeUnit unit) -423throws InterruptedException, ExecutionException, TimeoutException { -424 if (!this.isAbortInProgress) { -425return false; -426 } -427 super.get(timeout, unit); -428 return true; -429} -430 } -431 -432 /** @return Connection used by this object. */ -433 @Override -434 public Connection getConnection() { -435return connection; -436 } -437 -438 @Override -439 public boolean tableExists(final TableName tableName) throws IOException { -440return executeCallable(new RpcRetryingCallableBoolean() { -441 @Override -442 protected Boolean rpcCall(int callTimeout) throws Exception { -443return MetaTableAccessor.tableExists(connection, tableName); -444 } -445}); -446 } -447 -448 @Override -449 public HTableDescriptor[] listTables() throws IOException { -450return listTables((Pattern)null, false); -451 } -452 -453 @Override -454 public HTableDescriptor[] listTables(Pattern pattern) throws IOException {
[27/51] [partial] hbase-site git commit: Published site at 9101fc246f86445006bfbcdfda5cc495016dc280.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/65565d77/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html b/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html index a9d814d..e65cf97 100644 --- a/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html +++ b/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -288,13 +288,13 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html private https://docs.oracle.com/javase/8/docs/api/java/lang/ThreadGroup.html?is-external=true; title="class or interface in java.lang">ThreadGroup threadGroup -Created in the start(int, boolean) method. +Created in the #start(int, boolean) method. private TimeoutExecutorThread timeoutExecutor -Created in the start(int, boolean) method. +Created in the #start(int, boolean) method. @@ -308,7 +308,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html private https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CopyOnWriteArrayList.html?is-external=true; title="class or interface in java.util.concurrent">CopyOnWriteArrayListProcedureExecutor.WorkerThread workerThreads -Created in the start(int, boolean) method. +Created in the #start(int, boolean) method. @@ -490,107 +490,114 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true; title="class or interface in java.lang">InterruptedExceptione) +void +init(intnumThreads, +booleanabortOnCorruption) +Initialize the procedure executor, but do not start workers. + + + private Procedure[] initializeChildren(RootProcedureStateprocStack, Procedureprocedure, Procedure[]subprocs) - + boolean isFinished(longprocId) Return true if the procedure is finished. - + boolean isProcedureOwner(longprocId, Useruser) Check if the user is this procedure's owner - + boolean isRunning() - + boolean isStarted(longprocId) Return true if the procedure is started. - + void join() - + private void load(booleanabortOnCorruption) - + private void loadProcedures(ProcedureStore.ProcedureIteratorprocIter, booleanabortOnCorruption) - + private long nextProcId() - + private Procedure prepareProcedure(Procedureproc) - + private void procedureFinished(Procedureproc) - + private long pushProcedure(Procedureproc) - + void refreshConfiguration(org.apache.hadoop.conf.Configurationconf) - + void registerListener(ProcedureExecutor.ProcedureExecutorListenerlistener) - + long registerNonce(NonceKeynonceKey) Register a nonce for a procedure that is going to be submitted. - + private void releaseLock(Procedureproc, booleanforce) - + boolean removeChore(ProcedureInMemoryChorechore) Remove a chore procedure from the executor - + void removeResult(longprocId) Mark the specified completed procedure, as ready to remove. - + private void sendProcedureAddedNotification(longprocId) - + private void sendProcedureFinishedNotification(longprocId) - + private void sendProcedureLoadedNotification(longprocId) - + void setFailureResultForNonce(NonceKeynonceKey, https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in
[27/51] [partial] hbase-site git commit: Published site at 0b28155d274910b4e667b949d51f78809a1eff0b.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e11cf2cb/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html index f236300..513d2ad 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html @@ -26,1048 +26,1115 @@ 018 019package org.apache.hadoop.hbase.backup.impl; 020 -021import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH; -022import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH_DESC; -023import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG; -024import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC; -025import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH; -026import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC; -027import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER; -028import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC; -029import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET; -030import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC; -031import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC; -032import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE; -033import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC; -034import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC; -035import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS; -036import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC; -037import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME; -038import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC; -039 -040import java.io.IOException; -041import java.net.URI; -042import java.util.List; +021import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BACKUP_LIST_DESC; +022import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH; +023import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH_DESC; +024import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG; +025import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC; +026import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_KEEP; +027import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_KEEP_DESC; +028import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_LIST; +029import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH; +030import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC; +031import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER; +032import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC; +033import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET; +034import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC; +035import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC; +036import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE; +037import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC; +038import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC; +039import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS; +040import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC; +041import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME; +042import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC; 043 -044import org.apache.commons.lang3.StringUtils; -045import org.apache.hadoop.conf.Configuration; -046import org.apache.hadoop.conf.Configured; -047import org.apache.hadoop.fs.FileSystem; -048import org.apache.hadoop.fs.Path; -049import org.apache.hadoop.hbase.HBaseConfiguration; -050import org.apache.hadoop.hbase.TableName; -051import org.apache.hadoop.hbase.backup.BackupAdmin; -052import org.apache.hadoop.hbase.backup.BackupInfo; -053import
[27/51] [partial] hbase-site git commit: Published site at 7d3750bd9fc9747623549c242cc4171e224b3eaf.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/3469cbc0/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html index 275ca84..b9a30c4 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html @@ -1499,782 +1499,786 @@ 1491 final RpcController controller, 1492 final IsInMaintenanceModeRequest request) throws ServiceException { 1493IsInMaintenanceModeResponse.Builder response = IsInMaintenanceModeResponse.newBuilder(); -1494 response.setInMaintenanceMode(master.isInMaintenanceMode()); -1495return response.build(); -1496 } -1497 -1498 @Override -1499 public UnassignRegionResponse unassignRegion(RpcController controller, -1500 UnassignRegionRequest req) throws ServiceException { -1501try { -1502 final byte [] regionName = req.getRegion().getValue().toByteArray(); -1503 RegionSpecifierType type = req.getRegion().getType(); -1504 final boolean force = req.getForce(); -1505 UnassignRegionResponse urr = UnassignRegionResponse.newBuilder().build(); -1506 -1507 master.checkInitialized(); -1508 if (type != RegionSpecifierType.REGION_NAME) { -1509LOG.warn("unassignRegion specifier type: expected: " + RegionSpecifierType.REGION_NAME -1510 + " actual: " + type); -1511 } -1512 PairRegionInfo, ServerName pair = -1513 MetaTableAccessor.getRegion(master.getConnection(), regionName); -1514 if (Bytes.equals(RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName(),regionName)) { -1515pair = new Pair(RegionInfoBuilder.FIRST_META_REGIONINFO, -1516 master.getMetaTableLocator().getMetaRegionLocation(master.getZooKeeper())); -1517 } -1518 if (pair == null) { -1519throw new UnknownRegionException(Bytes.toString(regionName)); -1520 } -1521 -1522 RegionInfo hri = pair.getFirst(); -1523 if (master.cpHost != null) { -1524master.cpHost.preUnassign(hri, force); -1525 } -1526 LOG.debug(master.getClientIdAuditPrefix() + " unassign " + hri.getRegionNameAsString() -1527 + " in current location if it is online and reassign.force=" + force); -1528 master.getAssignmentManager().unassign(hri); -1529 if (master.cpHost != null) { -1530master.cpHost.postUnassign(hri, force); -1531 } -1532 -1533 return urr; -1534} catch (IOException ioe) { -1535 throw new ServiceException(ioe); -1536} -1537 } -1538 -1539 @Override -1540 public ReportRegionStateTransitionResponse reportRegionStateTransition(RpcController c, -1541 ReportRegionStateTransitionRequest req) throws ServiceException { -1542try { -1543 master.checkServiceStarted(); -1544 return master.getAssignmentManager().reportRegionStateTransition(req); -1545} catch (IOException ioe) { -1546 throw new ServiceException(ioe); -1547} -1548 } -1549 -1550 @Override -1551 public SetQuotaResponse setQuota(RpcController c, SetQuotaRequest req) -1552 throws ServiceException { -1553try { -1554 master.checkInitialized(); -1555 return master.getMasterQuotaManager().setQuota(req); -1556} catch (Exception e) { -1557 throw new ServiceException(e); -1558} -1559 } -1560 -1561 @Override -1562 public MajorCompactionTimestampResponse getLastMajorCompactionTimestamp(RpcController controller, -1563 MajorCompactionTimestampRequest request) throws ServiceException { -1564 MajorCompactionTimestampResponse.Builder response = -1565 MajorCompactionTimestampResponse.newBuilder(); -1566try { -1567 master.checkInitialized(); -1568 response.setCompactionTimestamp(master.getLastMajorCompactionTimestamp(ProtobufUtil -1569 .toTableName(request.getTableName(; -1570} catch (IOException e) { -1571 throw new ServiceException(e); -1572} -1573return response.build(); -1574 } -1575 -1576 @Override -1577 public MajorCompactionTimestampResponse getLastMajorCompactionTimestampForRegion( -1578 RpcController controller, MajorCompactionTimestampForRegionRequest request) -1579 throws ServiceException { -1580 MajorCompactionTimestampResponse.Builder response = -1581 MajorCompactionTimestampResponse.newBuilder(); -1582try { -1583 master.checkInitialized(); -1584 response.setCompactionTimestamp(master.getLastMajorCompactionTimestampForRegion(request -1585 .getRegion().getValue().toByteArray())); -1586} catch (IOException e) { -1587 throw new
[27/51] [partial] hbase-site git commit: Published site at 997747076d8ec0b4346d7cb99c4b0667a7c14905.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4df09ed9/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html index 3da432b..d30fa8f 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html @@ -928,7690 +928,7698 @@ 920 CollectionHStore stores = this.stores.values(); 921 try { 922// update the stores that we are replaying -923 stores.forEach(HStore::startReplayingFromWAL); -924// Recover any edits if available. -925maxSeqId = Math.max(maxSeqId, -926 replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, status)); -927// Make sure mvcc is up to max. -928this.mvcc.advanceTo(maxSeqId); -929 } finally { -930// update the stores that we are done replaying -931 stores.forEach(HStore::stopReplayingFromWAL); -932 } -933} -934this.lastReplayedOpenRegionSeqId = maxSeqId; -935 -936 this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this)); -937this.writestate.flushRequested = false; -938this.writestate.compacting.set(0); -939 -940if (this.writestate.writesEnabled) { -941 // Remove temporary data left over from old regions -942 status.setStatus("Cleaning up temporary data from old regions"); -943 fs.cleanupTempDir(); -944} -945 -946if (this.writestate.writesEnabled) { -947 status.setStatus("Cleaning up detritus from prior splits"); -948 // Get rid of any splits or merges that were lost in-progress. Clean out -949 // these directories here on open. We may be opening a region that was -950 // being split but we crashed in the middle of it all. -951 fs.cleanupAnySplitDetritus(); -952 fs.cleanupMergesDir(); -953} -954 -955// Initialize split policy -956this.splitPolicy = RegionSplitPolicy.create(this, conf); -957 -958// Initialize flush policy -959this.flushPolicy = FlushPolicyFactory.create(this, conf); -960 -961long lastFlushTime = EnvironmentEdgeManager.currentTime(); -962for (HStore store: stores.values()) { -963 this.lastStoreFlushTimeMap.put(store, lastFlushTime); -964} -965 -966// Use maximum of log sequenceid or that which was found in stores -967// (particularly if no recovered edits, seqid will be -1). -968long maxSeqIdFromFile = -969 WALSplitter.getMaxRegionSequenceId(fs.getFileSystem(), fs.getRegionDir()); -970long nextSeqId = Math.max(maxSeqId, maxSeqIdFromFile) + 1; -971if (writestate.writesEnabled) { -972 WALSplitter.writeRegionSequenceIdFile(fs.getFileSystem(), fs.getRegionDir(), nextSeqId - 1); -973} -974 -975LOG.info("Opened {}; next sequenceid={}", this.getRegionInfo().getShortNameToLog(), nextSeqId); -976 -977// A region can be reopened if failed a split; reset flags -978this.closing.set(false); -979this.closed.set(false); -980 -981if (coprocessorHost != null) { -982 status.setStatus("Running coprocessor post-open hooks"); -983 coprocessorHost.postOpen(); -984} +923LOG.debug("replaying wal for " + this.getRegionInfo().getEncodedName()); +924 stores.forEach(HStore::startReplayingFromWAL); +925// Recover any edits if available. +926maxSeqId = Math.max(maxSeqId, +927 replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, status)); +928// Make sure mvcc is up to max. +929this.mvcc.advanceTo(maxSeqId); +930 } finally { +931LOG.debug("stopping wal replay for " + this.getRegionInfo().getEncodedName()); +932// update the stores that we are done replaying +933 stores.forEach(HStore::stopReplayingFromWAL); +934 } +935} +936this.lastReplayedOpenRegionSeqId = maxSeqId; +937 +938 this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this)); +939this.writestate.flushRequested = false; +940this.writestate.compacting.set(0); +941 +942if (this.writestate.writesEnabled) { +943 LOG.debug("Cleaning up temporary data for " + this.getRegionInfo().getEncodedName()); +944 // Remove temporary data left over from old regions +945 status.setStatus("Cleaning up temporary data from old regions"); +946 fs.cleanupTempDir(); +947} +948 +949if (this.writestate.writesEnabled) { +950 status.setStatus("Cleaning up detritus from prior splits"); +951 // Get rid of any splits or merges that were lost in-progress. Clean out +952 // these directories here on open.
[27/51] [partial] hbase-site git commit: Published site at f3d1c021de2264301f68eadb9ef126ff83d7ef53.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/883dde2f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html index 7d27402..3da432b 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html @@ -2560,8 +2560,8 @@ 2552} 2553MemStoreSize mss = this.memStoreSizing.getMemStoreSize(); 2554LOG.info("Flushing " + + storesToFlush.size() + "/" + stores.size() + " column families," + -2555" memstore data size=" + StringUtils.byteDesc(mss.getDataSize()) + -2556" memstore heap size=" + StringUtils.byteDesc(mss.getHeapSize()) + +2555" dataSize=" + StringUtils.byteDesc(mss.getDataSize()) + +2556" heapSize=" + StringUtils.byteDesc(mss.getHeapSize()) + 2557((perCfExtras != null perCfExtras.length() 0)? perCfExtras.toString(): "") + 2558((wal != null) ? "" : "; WAL is null, using passed sequenceid=" + sequenceId)); 2559 } @@ -2761,10 +2761,10 @@ 2753long time = EnvironmentEdgeManager.currentTime() - startTime; 2754MemStoreSize mss = prepareResult.totalFlushableSize.getMemStoreSize(); 2755long memstoresize = this.memStoreSizing.getMemStoreSize().getDataSize(); -2756String msg = "Finished memstore flush;" -2757+ " data size ~" + StringUtils.byteDesc(mss.getDataSize()) + "/" + mss.getDataSize() -2758+ ", heap size ~" + StringUtils.byteDesc(mss.getHeapSize()) + "/" + mss.getHeapSize() -2759+ ", currentsize=" + StringUtils.byteDesc(memstoresize) + "/" + memstoresize +2756String msg = "Finished flush of" +2757+ " dataSize ~" + StringUtils.byteDesc(mss.getDataSize()) + "/" + mss.getDataSize() +2758+ ", heapSize ~" + StringUtils.byteDesc(mss.getHeapSize()) + "/" + mss.getHeapSize() +2759+ ", currentSize=" + StringUtils.byteDesc(memstoresize) + "/" + memstoresize 2760+ " for " + this.getRegionInfo().getEncodedName() + " in " + time + "ms, sequenceid=" 2761+ flushOpSeqId + ", compaction requested=" + compactionRequested 2762+ ((wal == null) ? "; wal=null" : ""); @@ -4244,7 +4244,7 @@ 4236 * @param cellItr 4237 * @param now 4238 */ -4239 public void updateCellTimestamps(final IterableListCell cellItr, final byte[] now) +4239 private static void updateCellTimestamps(final IterableListCell cellItr, final byte[] now) 4240 throws IOException { 4241for (ListCell cells: cellItr) { 4242 if (cells == null) continue; @@ -4299,12 +4299,12 @@ 4291 requestFlush(); 4292 // Don't print current limit because it will vary too much. The message is used as a key 4293 // over in RetriesExhaustedWithDetailsException processing. -4294 throw new RegionTooBusyException("Over memstore limit; regionName=" + -4295 (this.getRegionInfo() == null? "unknown": this.getRegionInfo().getEncodedName()) + -4296 ", server=" + (this.getRegionServerServices() == null ? "unknown": -4297 this.getRegionServerServices().getServerName()) + -4298 ", blockingMemStoreSize=" + -4299 org.apache.hadoop.hbase.procedure2.util.StringUtils.humanSize(blockingMemStoreSize)); +4294 throw new RegionTooBusyException("Over memstore limit=" + +4295 org.apache.hadoop.hbase.procedure2.util.StringUtils.humanSize(this.blockingMemStoreSize) + +4296", regionName=" + +4297 (this.getRegionInfo() == null? "unknown": this.getRegionInfo().getEncodedName()) + +4298 ", server=" + (this.getRegionServerServices() == null? "unknown": +4299 this.getRegionServerServices().getServerName())); 4300} 4301 } 4302 @@ -8610,7 +8610,8 @@ 8602 public void requestFlush(FlushLifeCycleTracker tracker) throws IOException { 8603requestFlush0(tracker); 8604 } -8605} +8605 +8606} http://git-wip-us.apache.org/repos/asf/hbase-site/blob/883dde2f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html index 7d27402..3da432b 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html @@ -2560,8 +2560,8 @@ 2552} 2553MemStoreSize mss = this.memStoreSizing.getMemStoreSize(); 2554
[27/51] [partial] hbase-site git commit: Published site at cf529f18a9959589fa635f78df4840472526ea2c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7bcc960d/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html index a97dfdc..2b1b6c6 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html @@ -2370,1287 +2370,1292 @@ 2362 } 2363 2364 @Override -2365 public long modifyTable(final TableName tableName, final TableDescriptor descriptor, +2365 public long modifyTable(final TableName tableName, final TableDescriptor newDescriptor, 2366 final long nonceGroup, final long nonce) throws IOException { 2367checkInitialized(); -2368 sanityCheckTableDescriptor(descriptor); +2368 sanityCheckTableDescriptor(newDescriptor); 2369 2370return MasterProcedureUtil.submitProcedure( 2371new MasterProcedureUtil.NonceProcedureRunnable(this, nonceGroup, nonce) { 2372 @Override 2373 protected void run() throws IOException { -2374 getMaster().getMasterCoprocessorHost().preModifyTable(tableName, descriptor); -2375 -2376 LOG.info(getClientIdAuditPrefix() + " modify " + tableName); +2374TableDescriptor oldDescriptor = getMaster().getTableDescriptors().get(tableName); +2375 getMaster().getMasterCoprocessorHost() +2376 .preModifyTable(tableName, oldDescriptor, newDescriptor); 2377 -2378// Execute the operation synchronously - wait for the operation completes before continuing. -2379// -2380// We need to wait for the procedure to potentially fail due to "prepare" sanity -2381// checks. This will block only the beginning of the procedure. See HBASE-19953. -2382ProcedurePrepareLatch latch = ProcedurePrepareLatch.createBlockingLatch(); -2383submitProcedure(new ModifyTableProcedure(procedureExecutor.getEnvironment(), -2384descriptor, latch)); -2385latch.await(); -2386 -2387 getMaster().getMasterCoprocessorHost().postModifyTable(tableName, descriptor); -2388 } -2389 -2390 @Override -2391 protected String getDescription() { -2392return "ModifyTableProcedure"; -2393 } -2394}); -2395 } -2396 -2397 public long restoreSnapshot(final SnapshotDescription snapshotDesc, -2398 final long nonceGroup, final long nonce, final boolean restoreAcl) throws IOException { -2399checkInitialized(); -2400 getSnapshotManager().checkSnapshotSupport(); -2401 -2402// Ensure namespace exists. Will throw exception if non-known NS. -2403final TableName dstTable = TableName.valueOf(snapshotDesc.getTable()); -2404 getClusterSchema().getNamespace(dstTable.getNamespaceAsString()); -2405 -2406return MasterProcedureUtil.submitProcedure( -2407new MasterProcedureUtil.NonceProcedureRunnable(this, nonceGroup, nonce) { -2408 @Override -2409 protected void run() throws IOException { -2410 setProcId( -2411 getSnapshotManager().restoreOrCloneSnapshot(snapshotDesc, getNonceKey(), restoreAcl)); -2412 } -2413 -2414 @Override -2415 protected String getDescription() { -2416return "RestoreSnapshotProcedure"; -2417 } -2418}); -2419 } -2420 -2421 private void checkTableExists(final TableName tableName) -2422 throws IOException, TableNotFoundException { -2423if (!MetaTableAccessor.tableExists(getConnection(), tableName)) { -2424 throw new TableNotFoundException(tableName); -2425} -2426 } -2427 -2428 @Override -2429 public void checkTableModifiable(final TableName tableName) -2430 throws IOException, TableNotFoundException, TableNotDisabledException { -2431if (isCatalogTable(tableName)) { -2432 throw new IOException("Can't modify catalog tables"); -2433} -2434checkTableExists(tableName); -2435TableState ts = getTableStateManager().getTableState(tableName); -2436if (!ts.isDisabled()) { -2437 throw new TableNotDisabledException("Not DISABLED; " + ts); -2438} -2439 } -2440 -2441 public ClusterMetrics getClusterMetricsWithoutCoprocessor() throws InterruptedIOException { -2442return getClusterMetricsWithoutCoprocessor(EnumSet.allOf(Option.class)); -2443 } -2444 -2445 public ClusterMetrics getClusterMetricsWithoutCoprocessor(EnumSetOption options) -2446 throws InterruptedIOException { -2447ClusterMetricsBuilder builder = ClusterMetricsBuilder.newBuilder(); -2448// given that hbase1 can't submit the request with Option, -2449// we return all information to client if the list of Option is empty. -2450if (options.isEmpty()) { -2451 options = EnumSet.allOf(Option.class); -2452} -2453 -2454for (Option opt : options) { -2455 switch (opt) {
[27/51] [partial] hbase-site git commit: Published site at 021f66d11d2cbb7308308093e29e69d6e7661ee9.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/92a26cfb/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html index 4a879bb..7d27402 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.Visitor.html @@ -300,7 +300,7 @@ 292 private MapString, com.google.protobuf.Service coprocessorServiceHandlers = Maps.newHashMap(); 293 294 // Track data size in all memstores -295 private final MemStoreSizing memStoreSize = new MemStoreSizing(); +295 private final MemStoreSizing memStoreSizing = new ThreadSafeMemStoreSizing(); 296 private final RegionServicesForStores regionServicesForStores = new RegionServicesForStores(this); 297 298 // Debug possible data loss due to WAL off @@ -1218,7389 +1218,7399 @@ 1210 * Increase the size of mem store in this region and the size of global mem 1211 * store 1212 */ -1213 public void incMemStoreSize(MemStoreSize memStoreSize) { -1214if (this.rsAccounting != null) { -1215 rsAccounting.incGlobalMemStoreSize(memStoreSize); -1216} -1217long dataSize; -1218synchronized (this.memStoreSize) { -1219 this.memStoreSize.incMemStoreSize(memStoreSize); -1220 dataSize = this.memStoreSize.getDataSize(); -1221} -1222 checkNegativeMemStoreDataSize(dataSize, memStoreSize.getDataSize()); -1223 } -1224 -1225 public void decrMemStoreSize(MemStoreSize memStoreSize) { -1226if (this.rsAccounting != null) { -1227 rsAccounting.decGlobalMemStoreSize(memStoreSize); -1228} -1229long size; -1230synchronized (this.memStoreSize) { -1231 this.memStoreSize.decMemStoreSize(memStoreSize); -1232 size = this.memStoreSize.getDataSize(); +1213 void incMemStoreSize(MemStoreSize mss) { +1214incMemStoreSize(mss.getDataSize(), mss.getHeapSize(), mss.getOffHeapSize()); +1215 } +1216 +1217 void incMemStoreSize(long dataSizeDelta, long heapSizeDelta, long offHeapSizeDelta) { +1218if (this.rsAccounting != null) { +1219 rsAccounting.incGlobalMemStoreSize(dataSizeDelta, heapSizeDelta, offHeapSizeDelta); +1220} +1221long dataSize = +1222 this.memStoreSizing.incMemStoreSize(dataSizeDelta, heapSizeDelta, offHeapSizeDelta); +1223 checkNegativeMemStoreDataSize(dataSize, dataSizeDelta); +1224 } +1225 +1226 void decrMemStoreSize(MemStoreSize mss) { +1227decrMemStoreSize(mss.getDataSize(), mss.getHeapSize(), mss.getOffHeapSize()); +1228 } +1229 +1230 void decrMemStoreSize(long dataSizeDelta, long heapSizeDelta, long offHeapSizeDelta) { +1231if (this.rsAccounting != null) { +1232 rsAccounting.decGlobalMemStoreSize(dataSizeDelta, heapSizeDelta, offHeapSizeDelta); 1233} -1234checkNegativeMemStoreDataSize(size, -memStoreSize.getDataSize()); -1235 } -1236 -1237 private void checkNegativeMemStoreDataSize(long memStoreDataSize, long delta) { -1238// This is extremely bad if we make memStoreSize negative. Log as much info on the offending -1239// caller as possible. (memStoreSize might be a negative value already -- freeing memory) -1240if (memStoreDataSize 0) { -1241 LOG.error("Asked to modify this region's (" + this.toString() -1242 + ") memStoreSize to a negative value which is incorrect. Current memStoreSize=" -1243 + (memStoreDataSize - delta) + ", delta=" + delta, new Exception()); -1244} -1245 } -1246 -1247 @Override -1248 public RegionInfo getRegionInfo() { -1249return this.fs.getRegionInfo(); -1250 } -1251 -1252 /** -1253 * @return Instance of {@link RegionServerServices} used by this HRegion. -1254 * Can be null. -1255 */ -1256 RegionServerServices getRegionServerServices() { -1257return this.rsServices; -1258 } -1259 -1260 @Override -1261 public long getReadRequestsCount() { -1262return readRequestsCount.sum(); -1263 } -1264 -1265 @Override -1266 public long getFilteredReadRequestsCount() { -1267return filteredReadRequestsCount.sum(); -1268 } -1269 -1270 @Override -1271 public long getWriteRequestsCount() { -1272return writeRequestsCount.sum(); -1273 } -1274 -1275 @Override -1276 public long getMemStoreDataSize() { -1277return memStoreSize.getDataSize(); -1278 } -1279 -1280 @Override -1281 public long getMemStoreHeapSize() { -1282return memStoreSize.getHeapSize(); -1283 } -1284 -1285 @Override -1286 public long getMemStoreOffHeapSize() { -1287return memStoreSize.getOffHeapSize(); -1288 } -1289 -1290 /** @return store services for this region, to access services required by store level
[27/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import
[27/51] [partial] hbase-site git commit: Published site at 87f5b5f3411d96c31b4cb61b9a57ced22be91d1f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/de18d468/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.MetaEntry.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.MetaEntry.html b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.MetaEntry.html index e1bc325..63e7421 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.MetaEntry.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.MetaEntry.html @@ -66,5125 +66,5224 @@ 058import java.util.concurrent.TimeoutException; 059import java.util.concurrent.atomic.AtomicBoolean; 060import java.util.concurrent.atomic.AtomicInteger; -061import org.apache.commons.io.IOUtils; -062import org.apache.commons.lang3.RandomStringUtils; -063import org.apache.commons.lang3.StringUtils; -064import org.apache.hadoop.conf.Configuration; -065import org.apache.hadoop.conf.Configured; -066import org.apache.hadoop.fs.FSDataOutputStream; -067import org.apache.hadoop.fs.FileStatus; -068import org.apache.hadoop.fs.FileSystem; -069import org.apache.hadoop.fs.Path; -070import org.apache.hadoop.fs.permission.FsAction; -071import org.apache.hadoop.fs.permission.FsPermission; -072import org.apache.hadoop.hbase.Abortable; -073import org.apache.hadoop.hbase.Cell; -074import org.apache.hadoop.hbase.CellUtil; -075import org.apache.hadoop.hbase.ClusterMetrics; -076import org.apache.hadoop.hbase.ClusterMetrics.Option; -077import org.apache.hadoop.hbase.HBaseConfiguration; -078import org.apache.hadoop.hbase.HBaseInterfaceAudience; -079import org.apache.hadoop.hbase.HConstants; -080import org.apache.hadoop.hbase.HRegionInfo; -081import org.apache.hadoop.hbase.HRegionLocation; -082import org.apache.hadoop.hbase.KeyValue; -083import org.apache.hadoop.hbase.MasterNotRunningException; -084import org.apache.hadoop.hbase.MetaTableAccessor; -085import org.apache.hadoop.hbase.RegionLocations; -086import org.apache.hadoop.hbase.ServerName; -087import org.apache.hadoop.hbase.TableName; -088import org.apache.hadoop.hbase.ZooKeeperConnectionException; -089import org.apache.hadoop.hbase.client.Admin; -090import org.apache.hadoop.hbase.client.ClusterConnection; -091import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor; -092import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder; -093import org.apache.hadoop.hbase.client.Connection; -094import org.apache.hadoop.hbase.client.ConnectionFactory; -095import org.apache.hadoop.hbase.client.Delete; -096import org.apache.hadoop.hbase.client.Get; -097import org.apache.hadoop.hbase.client.Put; -098import org.apache.hadoop.hbase.client.RegionInfo; -099import org.apache.hadoop.hbase.client.RegionInfoBuilder; -100import org.apache.hadoop.hbase.client.RegionReplicaUtil; -101import org.apache.hadoop.hbase.client.Result; -102import org.apache.hadoop.hbase.client.RowMutations; -103import org.apache.hadoop.hbase.client.Table; -104import org.apache.hadoop.hbase.client.TableDescriptor; -105import org.apache.hadoop.hbase.client.TableDescriptorBuilder; -106import org.apache.hadoop.hbase.client.TableState; -107import org.apache.hadoop.hbase.io.FileLink; -108import org.apache.hadoop.hbase.io.HFileLink; -109import org.apache.hadoop.hbase.io.hfile.CacheConfig; -110import org.apache.hadoop.hbase.io.hfile.HFile; -111import org.apache.hadoop.hbase.log.HBaseMarkers; -112import org.apache.hadoop.hbase.master.MasterFileSystem; -113import org.apache.hadoop.hbase.master.RegionState; -114import org.apache.hadoop.hbase.regionserver.HRegion; -115import org.apache.hadoop.hbase.regionserver.HRegionFileSystem; -116import org.apache.hadoop.hbase.regionserver.StoreFileInfo; -117import org.apache.hadoop.hbase.replication.ReplicationException; -118import org.apache.hadoop.hbase.security.AccessDeniedException; -119import org.apache.hadoop.hbase.security.UserProvider; -120import org.apache.hadoop.hbase.util.Bytes.ByteArrayComparator; -121import org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE; -122import org.apache.hadoop.hbase.util.hbck.HFileCorruptionChecker; -123import org.apache.hadoop.hbase.util.hbck.ReplicationChecker; -124import org.apache.hadoop.hbase.util.hbck.TableIntegrityErrorHandler; -125import org.apache.hadoop.hbase.util.hbck.TableIntegrityErrorHandlerImpl; -126import org.apache.hadoop.hbase.wal.WAL; -127import org.apache.hadoop.hbase.wal.WALFactory; -128import org.apache.hadoop.hbase.wal.WALSplitter; -129import org.apache.hadoop.hbase.zookeeper.MetaTableLocator; -130import org.apache.hadoop.hbase.zookeeper.ZKUtil; -131import org.apache.hadoop.hbase.zookeeper.ZKWatcher; -132import org.apache.hadoop.hbase.zookeeper.ZNodePaths; -133import org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException; -134import org.apache.hadoop.ipc.RemoteException; -135import org.apache.hadoop.security.UserGroupInformation; -136import
[27/51] [partial] hbase-site git commit: Published site at 2912c953551bedbfbf30c32c156ed7bb187d54c3.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d220bc5e/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/RowCounter.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/RowCounter.html b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/RowCounter.html index 56ec3ac..8f4cacb 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/RowCounter.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/RowCounter.html @@ -229,9 +229,9 @@ 221 * Note that we don't document --expected-count, because it's intended for test. 222 */ 223 private static void printUsage() { -224System.err.println("Usage: RowCounter [options] tablename " + -225"[--starttime=[start] --endtime=[end] " + -226 "[--range=[startKey],[endKey][;[startKey],[endKey]...]] [column1 column2...]"); +224System.err.println("Usage: hbase rowcounter [options] tablename " +225+ "[--starttime=start --endtime=end] " +226+ "[--range=[startKey],[endKey][;[startKey],[endKey]...]] [column1 column2...]"); 227System.err.println("For performance consider the following options:\n" 228+ "-Dhbase.client.scanner.caching=100\n" 229+ "-Dmapreduce.map.speculative=false"); http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d220bc5e/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.WALKeyValueMapper.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.WALKeyValueMapper.html b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.WALKeyValueMapper.html index a5477ac..9501e97 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.WALKeyValueMapper.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.WALKeyValueMapper.html @@ -357,46 +357,51 @@ 349 System.err.println("ERROR: " + errorMsg); 350} 351System.err.println("Usage: " + NAME + " [options] wal inputdir tables [tableMappings]"); -352System.err.println("Read all WAL entries for tables."); -353System.err.println("If no tables (\"\") are specific, all tables are imported."); -354System.err.println("(Careful, even hbase:meta entries will be imported" + " in that case.)"); -355System.err.println("Otherwise tables is a comma separated list of tables.\n"); -356System.err.println("The WAL entries can be mapped to new set of tables via tableMapping."); -357 System.err.println("tableMapping is a command separated list of targettables."); +352System.err.println("Replay all WAL files into HBase."); +353System.err.println("tables is a comma separated list of tables."); +354System.err.println("If no tables (\"\") are specified, all tables are imported."); +355System.err.println("(Be careful, hbase:meta entries will be imported in this case.)\n"); +356System.err.println("WAL entries can be mapped to new set of tables via tableMappings."); +357 System.err.println("tableMappings is a comma separated list of target tables."); 358System.err.println("If specified, each table in tables must have a mapping.\n"); 359System.err.println("By default " + NAME + " will load data directly into HBase."); -360System.err.println("To generate HFiles for a bulk data load instead, pass the option:"); +360System.err.println("To generate HFiles for a bulk data load instead, pass the following option:"); 361System.err.println(" -D" + BULK_OUTPUT_CONF_KEY + "=/path/for/output"); 362System.err.println(" (Only one table can be specified, and no mapping is allowed!)"); -363System.err.println("Other options: (specify time range to WAL edit to consider)"); +363System.err.println("Time range options:"); 364System.err.println(" -D" + WALInputFormat.START_TIME_KEY + "=[date|ms]"); 365System.err.println(" -D" + WALInputFormat.END_TIME_KEY + "=[date|ms]"); -366System.err.println(" -D " + JOB_NAME_CONF_KEY -367+ "=jobName - use the specified mapreduce job name for the wal player"); -368System.err.println("For performance also consider the following options:\n" -369+ " -Dmapreduce.map.speculative=false\n" + " -Dmapreduce.reduce.speculative=false"); -370 } -371 -372 /** -373 * Main entry point. -374 * @param args The command line parameters. -375 * @throws Exception When running the job fails. -376 */ -377 public static void main(String[] args) throws Exception { -378int ret = ToolRunner.run(new WALPlayer(HBaseConfiguration.create()), args); -379System.exit(ret); -380 } -381 -382 @Override -383 public int run(String[] args) throws Exception { -384if (args.length 2) { -385 usage("Wrong number of arguments: " + args.length); -386 System.exit(-1); -387} -388Job job =
[27/51] [partial] hbase-site git commit: Published site at 2a2258656b2fcd92b967131b6c1f037363553bc4.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/org/apache/hadoop/hbase/util/RegionMover.MoveWithoutAck.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/util/RegionMover.MoveWithoutAck.html b/devapidocs/org/apache/hadoop/hbase/util/RegionMover.MoveWithoutAck.html index efc1791..6e81af2 100644 --- a/devapidocs/org/apache/hadoop/hbase/util/RegionMover.MoveWithoutAck.html +++ b/devapidocs/org/apache/hadoop/hbase/util/RegionMover.MoveWithoutAck.html @@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab"; -private static class RegionMover.MoveWithoutAck +private static class RegionMover.MoveWithoutAck extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/Callable.html?is-external=true; title="class or interface in java.util.concurrent">Callablehttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean Move Regions without Acknowledging.Usefule in case of RS shutdown as we might want to shut the @@ -228,7 +228,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre admin -privateAdmin admin +privateAdmin admin @@ -237,7 +237,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre region -privateRegionInfo region +privateRegionInfo region @@ -246,7 +246,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre targetServer -privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String targetServer +privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String targetServer @@ -255,7 +255,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre movedRegions -privatehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo movedRegions +privatehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo movedRegions @@ -264,7 +264,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre sourceServer -privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String sourceServer +privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String sourceServer @@ -281,7 +281,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre MoveWithoutAck -publicMoveWithoutAck(Adminadmin, +publicMoveWithoutAck(Adminadmin, RegionInforegionInfo, https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringsourceServer, https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringtargetServer, @@ -302,7 +302,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurre call -publichttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Booleancall() +publichttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Booleancall() Specified by: https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/Callable.html?is-external=true#call--; title="class or interface in java.util.concurrent">callin interfacehttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/Callable.html?is-external=true; title="class or interface in java.util.concurrent">Callablehttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/org/apache/hadoop/hbase/util/RegionMover.RegionMoverBuilder.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/util/RegionMover.RegionMoverBuilder.html b/devapidocs/org/apache/hadoop/hbase/util/RegionMover.RegionMoverBuilder.html index c451701..d4d8db5 100644 --- a/devapidocs/org/apache/hadoop/hbase/util/RegionMover.RegionMoverBuilder.html +++ b/devapidocs/org/apache/hadoop/hbase/util/RegionMover.RegionMoverBuilder.html @@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab"; -public static class RegionMover.RegionMoverBuilder +public static class
[27/51] [partial] hbase-site git commit: Published site at e468b4022f76688851b3e0c34722f01a56bd624f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/16541468/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html b/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html index dbb2f3e..c887b35 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html @@ -36,843 +36,849 @@ 028import org.apache.hadoop.hbase.CompareOperator; 029import org.apache.hadoop.hbase.HTableDescriptor; 030import org.apache.hadoop.hbase.TableName; -031import org.apache.yetus.audience.InterfaceAudience; -032import org.apache.hadoop.hbase.client.coprocessor.Batch; -033import org.apache.hadoop.hbase.filter.CompareFilter; -034import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel; -035 -036import com.google.protobuf.Descriptors; -037import com.google.protobuf.Message; -038import com.google.protobuf.Service; -039import com.google.protobuf.ServiceException; -040 -041/** -042 * Used to communicate with a single HBase table. -043 * Obtain an instance from a {@link Connection} and call {@link #close()} afterwards. -044 * -045 * pTable can be used to get, put, delete or scan data from a table. -046 * @see ConnectionFactory -047 * @see Connection -048 * @see Admin -049 * @see RegionLocator -050 * @since 0.99.0 -051 */ -052@InterfaceAudience.Public -053public interface Table extends Closeable { -054 /** -055 * Gets the fully qualified table name instance of this table. -056 */ -057 TableName getName(); -058 -059 /** -060 * Returns the {@link org.apache.hadoop.conf.Configuration} object used by this instance. -061 * p -062 * The reference returned is not a copy, so any change made to it will -063 * affect this instance. -064 */ -065 Configuration getConfiguration(); -066 -067 /** -068 * Gets the {@link org.apache.hadoop.hbase.HTableDescriptor table descriptor} for this table. -069 * @throws java.io.IOException if a remote or network exception occurs. -070 * @deprecated since 2.0 version and will be removed in 3.0 version. -071 * use {@link #getDescriptor()} -072 */ -073 @Deprecated -074 HTableDescriptor getTableDescriptor() throws IOException; -075 -076 /** -077 * Gets the {@link org.apache.hadoop.hbase.client.TableDescriptor table descriptor} for this table. -078 * @throws java.io.IOException if a remote or network exception occurs. -079 */ -080 TableDescriptor getDescriptor() throws IOException; -081 -082 /** -083 * Test for the existence of columns in the table, as specified by the Get. -084 * p -085 * -086 * This will return true if the Get matches one or more keys, false if not. -087 * p -088 * -089 * This is a server-side call so it prevents any data from being transfered to -090 * the client. -091 * -092 * @param get the Get -093 * @return true if the specified Get matches one or more keys, false if not -094 * @throws IOException e -095 */ -096 boolean exists(Get get) throws IOException; -097 -098 /** -099 * Test for the existence of columns in the table, as specified by the Gets. -100 * p -101 * -102 * This will return an array of booleans. Each value will be true if the related Get matches -103 * one or more keys, false if not. -104 * p -105 * -106 * This is a server-side call so it prevents any data from being transferred to -107 * the client. -108 * -109 * @param gets the Gets -110 * @return Array of boolean. True if the specified Get matches one or more keys, false if not. -111 * @throws IOException e -112 */ -113 boolean[] exists(ListGet gets) throws IOException; -114 -115 /** -116 * Test for the existence of columns in the table, as specified by the Gets. -117 * This will return an array of booleans. Each value will be true if the related Get matches -118 * one or more keys, false if not. -119 * This is a server-side call so it prevents any data from being transferred to -120 * the client. -121 * -122 * @param gets the Gets -123 * @return Array of boolean. True if the specified Get matches one or more keys, false if not. -124 * @throws IOException e -125 * @deprecated since 2.0 version and will be removed in 3.0 version. -126 * use {@link #exists(List)} -127 */ -128 @Deprecated -129 default boolean[] existsAll(ListGet gets) throws IOException { -130return exists(gets); -131 } -132 -133 /** -134 * Method that does a batch call on Deletes, Gets, Puts, Increments, Appends, RowMutations. -135 * The ordering of execution of the actions is not defined. Meaning if you do a Put and a -136 * Get in the same {@link #batch} call, you will not necessarily be -137 * guaranteed that the Get returns what the Put had put. -138 * -139 * @param actions list of Get, Put, Delete, Increment, Append,
[27/51] [partial] hbase-site git commit: Published site at 64061f896fe21512504e3886a400759e88b519da.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f1ebf5b6/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html index 5360e0a..d4296e6 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html @@ -627,13 +627,18 @@ service. https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true; title="class or interface in java.util.concurrent">ExecutorServicepool) +static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListPairhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">Long +MetaTableAccessor.getTableEncodedRegionNameAndLastBarrier(Connectionconn, + TableNametableName) + + static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo MetaTableAccessor.getTableRegions(Connectionconnection, TableNametableName) Gets all of the regions of the specified table. - + static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo MetaTableAccessor.getTableRegions(Connectionconnection, TableNametableName, @@ -641,14 +646,14 @@ service. Gets all of the regions of the specified table. - + static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListPairRegionInfo,ServerName MetaTableAccessor.getTableRegionsAndLocations(Connectionconnection, TableNametableName) Do not use this method to get meta table regions, use methods in MetaTableLocator instead. - + static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListPairRegionInfo,ServerName MetaTableAccessor.getTableRegionsAndLocations(Connectionconnection, TableNametableName, @@ -656,37 +661,37 @@ service. Do not use this method to get meta table regions, use methods in MetaTableLocator instead. - + static byte[] MetaTableAccessor.getTableStartRowForMeta(TableNametableName, MetaTableAccessor.QueryTypetype) - + static https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">OptionalTableState AsyncMetaTableAccessor.getTableState(AsyncTable?metaTable, TableNametableName) - + static TableState MetaTableAccessor.getTableState(Connectionconn, TableNametableName) Fetch table state for given table from META table - + static byte[] MetaTableAccessor.getTableStopRowForMeta(TableNametableName, MetaTableAccessor.QueryTypetype) - + static boolean TableName.isMetaTableName(TableNametn) - + TableDescriptor TableDescriptors.remove(TableNametablename) - + static void MetaTableAccessor.scanMeta(Connectionconnection, MetaTableAccessor.Visitorvisitor, @@ -697,7 +702,7 @@ service. given row. - + static void MetaTableAccessor.scanMeta(Connectionconnection, TableNametable, @@ -705,25 +710,25 @@ service. intmaxRows, MetaTableAccessor.Visitorvisitor) - + static void MetaTableAccessor.scanMetaForTableRegions(Connectionconnection, MetaTableAccessor.Visitorvisitor, TableNametableName) - + static https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean AsyncMetaTableAccessor.tableExists(AsyncTable?metaTable, TableNametableName) - + static boolean MetaTableAccessor.tableExists(Connectionconnection, TableNametableName) Checks if the specified table exists. - + static void MetaTableAccessor.updateTableState(Connectionconn, TableNametableName, @@ -2052,119 +2057,119 @@ service. private TableName -SnapshotDescription.table +RegionCoprocessorRpcChannel.table private TableName
[27/51] [partial] hbase-site git commit: Published site at 4cb40e6d846ce1f28ffb40d388c9efb753197813.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4dc2a2e8/devapidocs/org/apache/hadoop/hbase/class-use/Size.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/Size.html b/devapidocs/org/apache/hadoop/hbase/class-use/Size.html index 3f1b032..e8d1010 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/Size.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/Size.html @@ -189,130 +189,130 @@ Size -RegionLoad.getBloomFilterSize() -Deprecated. - - - -Size RegionMetrics.getBloomFilterSize() - -Size -RegionMetricsBuilder.RegionMetricsImpl.getBloomFilterSize() - Size -ServerLoad.getMaxHeapSize() +RegionLoad.getBloomFilterSize() Deprecated. Size -ServerMetrics.getMaxHeapSize() +RegionMetricsBuilder.RegionMetricsImpl.getBloomFilterSize() Size -ServerMetricsBuilder.ServerMetricsImpl.getMaxHeapSize() +ServerMetrics.getMaxHeapSize() Size -RegionLoad.getMemStoreSize() +ServerLoad.getMaxHeapSize() Deprecated. Size -RegionMetrics.getMemStoreSize() +ServerMetricsBuilder.ServerMetricsImpl.getMaxHeapSize() Size -RegionMetricsBuilder.RegionMetricsImpl.getMemStoreSize() +RegionMetrics.getMemStoreSize() Size -RegionLoad.getStoreFileIndexSize() +RegionLoad.getMemStoreSize() Deprecated. Size +RegionMetricsBuilder.RegionMetricsImpl.getMemStoreSize() + + +Size RegionMetrics.getStoreFileIndexSize() TODO: why we pass the same value to different counters? Currently, the value from getStoreFileIndexSize() is same with getStoreFileRootLevelIndexSize() see HRegionServer#createRegionLoad. - -Size -RegionMetricsBuilder.RegionMetricsImpl.getStoreFileIndexSize() - Size -RegionLoad.getStoreFileRootLevelIndexSize() +RegionLoad.getStoreFileIndexSize() Deprecated. Size -RegionMetrics.getStoreFileRootLevelIndexSize() +RegionMetricsBuilder.RegionMetricsImpl.getStoreFileIndexSize() Size -RegionMetricsBuilder.RegionMetricsImpl.getStoreFileRootLevelIndexSize() +RegionMetrics.getStoreFileRootLevelIndexSize() Size -RegionLoad.getStoreFileSize() +RegionLoad.getStoreFileRootLevelIndexSize() Deprecated. Size -RegionMetrics.getStoreFileSize() +RegionMetricsBuilder.RegionMetricsImpl.getStoreFileRootLevelIndexSize() Size -RegionMetricsBuilder.RegionMetricsImpl.getStoreFileSize() +RegionMetrics.getStoreFileSize() Size -RegionLoad.getStoreFileUncompressedDataIndexSize() +RegionLoad.getStoreFileSize() Deprecated. Size -RegionMetrics.getStoreFileUncompressedDataIndexSize() +RegionMetricsBuilder.RegionMetricsImpl.getStoreFileSize() Size -RegionMetricsBuilder.RegionMetricsImpl.getStoreFileUncompressedDataIndexSize() +RegionMetrics.getStoreFileUncompressedDataIndexSize() Size -RegionLoad.getUncompressedStoreFileSize() +RegionLoad.getStoreFileUncompressedDataIndexSize() Deprecated. Size -RegionMetrics.getUncompressedStoreFileSize() +RegionMetricsBuilder.RegionMetricsImpl.getStoreFileUncompressedDataIndexSize() Size -RegionMetricsBuilder.RegionMetricsImpl.getUncompressedStoreFileSize() +RegionMetrics.getUncompressedStoreFileSize() Size -ServerLoad.getUsedHeapSize() +RegionLoad.getUncompressedStoreFileSize() Deprecated. Size +RegionMetricsBuilder.RegionMetricsImpl.getUncompressedStoreFileSize() + + +Size ServerMetrics.getUsedHeapSize() + +Size +ServerLoad.getUsedHeapSize() +Deprecated. + + Size ServerMetricsBuilder.ServerMetricsImpl.getUsedHeapSize() http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4dc2a2e8/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html b/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html index 63833f7..72d579d 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableDescriptors.html @@ -122,11 +122,11 @@ TableDescriptors -HMaster.getTableDescriptors() +MasterServices.getTableDescriptors() TableDescriptors -MasterServices.getTableDescriptors() +HMaster.getTableDescriptors()
[27/51] [partial] hbase-site git commit: Published site at 8ab7b20f48951d77945181024f5e15842bc253c4.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6eb695c8/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html index ecf500c..0cd5a4e 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html @@ -238,8355 +238,8368 @@ 230 public static final String HBASE_MAX_CELL_SIZE_KEY = "hbase.server.keyvalue.maxsize"; 231 public static final int DEFAULT_MAX_CELL_SIZE = 10485760; 232 -233 public static final String HBASE_REGIONSERVER_MINIBATCH_SIZE = -234 "hbase.regionserver.minibatch.size"; -235 public static final int DEFAULT_HBASE_REGIONSERVER_MINIBATCH_SIZE = 2; -236 -237 /** -238 * This is the global default value for durability. All tables/mutations not -239 * defining a durability or using USE_DEFAULT will default to this value. -240 */ -241 private static final Durability DEFAULT_DURABILITY = Durability.SYNC_WAL; +233 /** +234 * This is the global default value for durability. All tables/mutations not +235 * defining a durability or using USE_DEFAULT will default to this value. +236 */ +237 private static final Durability DEFAULT_DURABILITY = Durability.SYNC_WAL; +238 +239 public static final String HBASE_REGIONSERVER_MINIBATCH_SIZE = +240 "hbase.regionserver.minibatch.size"; +241 public static final int DEFAULT_HBASE_REGIONSERVER_MINIBATCH_SIZE = 2; 242 -243 final AtomicBoolean closed = new AtomicBoolean(false); -244 -245 /* Closing can take some time; use the closing flag if there is stuff we don't -246 * want to do while in closing state; e.g. like offer this region up to the -247 * master as a region to close if the carrying regionserver is overloaded. -248 * Once set, it is never cleared. -249 */ -250 final AtomicBoolean closing = new AtomicBoolean(false); -251 -252 /** -253 * The max sequence id of flushed data on this region. There is no edit in memory that is -254 * less that this sequence id. -255 */ -256 private volatile long maxFlushedSeqId = HConstants.NO_SEQNUM; -257 -258 /** -259 * Record the sequence id of last flush operation. Can be in advance of -260 * {@link #maxFlushedSeqId} when flushing a single column family. In this case, -261 * {@link #maxFlushedSeqId} will be older than the oldest edit in memory. -262 */ -263 private volatile long lastFlushOpSeqId = HConstants.NO_SEQNUM; -264 -265 /** -266 * The sequence id of the last replayed open region event from the primary region. This is used -267 * to skip entries before this due to the possibility of replay edits coming out of order from -268 * replication. -269 */ -270 protected volatile long lastReplayedOpenRegionSeqId = -1L; -271 protected volatile long lastReplayedCompactionSeqId = -1L; -272 -273 // -274 // Members -275 // -276 -277 // map from a locked row to the context for that lock including: -278 // - CountDownLatch for threads waiting on that row -279 // - the thread that owns the lock (allow reentrancy) -280 // - reference count of (reentrant) locks held by the thread -281 // - the row itself -282 private final ConcurrentHashMapHashedBytes, RowLockContext lockedRows = -283 new ConcurrentHashMap(); -284 -285 protected final Mapbyte[], HStore stores = -286 new ConcurrentSkipListMap(Bytes.BYTES_RAWCOMPARATOR); +243 public static final String WAL_HSYNC_CONF_KEY = "hbase.wal.hsync"; +244 public static final boolean DEFAULT_WAL_HSYNC = false; +245 +246 final AtomicBoolean closed = new AtomicBoolean(false); +247 +248 /* Closing can take some time; use the closing flag if there is stuff we don't +249 * want to do while in closing state; e.g. like offer this region up to the +250 * master as a region to close if the carrying regionserver is overloaded. +251 * Once set, it is never cleared. +252 */ +253 final AtomicBoolean closing = new AtomicBoolean(false); +254 +255 /** +256 * The max sequence id of flushed data on this region. There is no edit in memory that is +257 * less that this sequence id. +258 */ +259 private volatile long maxFlushedSeqId = HConstants.NO_SEQNUM; +260 +261 /** +262 * Record the sequence id of last flush operation. Can be in advance of +263 * {@link #maxFlushedSeqId} when flushing a single column family. In this case, +264 * {@link #maxFlushedSeqId} will be older than the oldest edit in memory. +265 */ +266 private volatile
[27/51] [partial] hbase-site git commit: Published site at 00095a2ef9442e3fd86c04876c9d91f2f8b23ad8.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bd675fa3/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.StoreToFile.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.StoreToFile.html b/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.StoreToFile.html index f47d627..c3d225c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.StoreToFile.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/backup/HFileArchiver.StoreToFile.html @@ -117,219 +117,219 @@ 109 */ 110 public static boolean archiveRegion(FileSystem fs, Path rootdir, Path tableDir, Path regionDir) 111 throws IOException { -112if (LOG.isDebugEnabled()) { -113 LOG.debug("ARCHIVING " + regionDir.toString()); -114} -115 -116// otherwise, we archive the files -117// make sure we can archive -118if (tableDir == null || regionDir == null) { -119 LOG.error("No archive directory could be found because tabledir (" + tableDir -120 + ") or regiondir (" + regionDir + "was null. Deleting files instead."); -121 deleteRegionWithoutArchiving(fs, regionDir); -122 // we should have archived, but failed to. Doesn't matter if we deleted -123 // the archived files correctly or not. -124 return false; -125} -126 -127// make sure the regiondir lives under the tabledir -128 Preconditions.checkArgument(regionDir.toString().startsWith(tableDir.toString())); -129Path regionArchiveDir = HFileArchiveUtil.getRegionArchiveDir(rootdir, -130FSUtils.getTableName(tableDir), -131regionDir.getName()); -132 -133FileStatusConverter getAsFile = new FileStatusConverter(fs); -134// otherwise, we attempt to archive the store files -135 -136// build collection of just the store directories to archive -137CollectionFile toArchive = new ArrayList(); -138final PathFilter dirFilter = new FSUtils.DirFilter(fs); -139PathFilter nonHidden = new PathFilter() { -140 @Override -141 public boolean accept(Path file) { -142return dirFilter.accept(file) !file.getName().toString().startsWith("."); -143 } -144}; -145FileStatus[] storeDirs = FSUtils.listStatus(fs, regionDir, nonHidden); -146// if there no files, we can just delete the directory and return; -147if (storeDirs == null) { -148 LOG.debug("Region directory " + regionDir + " empty."); -149 return deleteRegionWithoutArchiving(fs, regionDir); -150} -151 -152// convert the files in the region to a File -153 toArchive.addAll(Lists.transform(Arrays.asList(storeDirs), getAsFile)); -154LOG.debug("Archiving " + toArchive); -155ListFile failedArchive = resolveAndArchive(fs, regionArchiveDir, toArchive, -156 EnvironmentEdgeManager.currentTime()); -157if (!failedArchive.isEmpty()) { -158 throw new FailedArchiveException("Failed to archive/delete all the files for region:" -159 + regionDir.getName() + " into " + regionArchiveDir -160 + ". Something is probably awry on the filesystem.", -161 Collections2.transform(failedArchive, FUNC_FILE_TO_PATH)); -162} -163// if that was successful, then we delete the region -164return deleteRegionWithoutArchiving(fs, regionDir); -165 } -166 -167 /** -168 * Remove from the specified region the store files of the specified column family, -169 * either by archiving them or outright deletion -170 * @param fs the filesystem where the store files live -171 * @param conf {@link Configuration} to examine to determine the archive directory -172 * @param parent Parent region hosting the store files -173 * @param tableDir {@link Path} to where the table is being stored (for building the archive path) -174 * @param family the family hosting the store files -175 * @throws IOException if the files could not be correctly disposed. -176 */ -177 public static void archiveFamily(FileSystem fs, Configuration conf, -178 RegionInfo parent, Path tableDir, byte[] family) throws IOException { -179Path familyDir = new Path(tableDir, new Path(parent.getEncodedName(), Bytes.toString(family))); -180archiveFamilyByFamilyDir(fs, conf, parent, familyDir, family); -181 } -182 -183 /** -184 * Removes from the specified region the store files of the specified column family, -185 * either by archiving them or outright deletion -186 * @param fs the filesystem where the store files live -187 * @param conf {@link Configuration} to examine to determine the archive directory -188 * @param parent Parent region hosting the store files -189 * @param familyDir {@link Path} to where the family is being stored -190 * @param family the family hosting the store files -191 * @throws IOException if the files could not be correctly disposed. -192 */
[27/51] [partial] hbase-site git commit: Published site at 22f4def942f8a3367d0ca6598317e9b9a7d0cfcd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8b1eaec1/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionEnvironment.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionEnvironment.html b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionEnvironment.html index 067b5c4..beb9075 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionEnvironment.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionEnvironment.html @@ -126,7 +126,7 @@ var activeTableTab = "activeTableTab"; -private static class RegionCoprocessorHost.RegionEnvironment +private static class RegionCoprocessorHost.RegionEnvironment extends BaseEnvironmentRegionCoprocessor implements RegionCoprocessorEnvironment Encapsulation of the environment of each coprocessor @@ -303,7 +303,7 @@ implements region -privateRegion region +privateRegion region @@ -312,7 +312,7 @@ implements sharedData -https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true; title="class or interface in java.util.concurrent">ConcurrentMaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object sharedData +https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true; title="class or interface in java.util.concurrent">ConcurrentMaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object sharedData @@ -321,7 +321,7 @@ implements metricRegistry -private finalMetricRegistry metricRegistry +private finalMetricRegistry metricRegistry @@ -330,7 +330,7 @@ implements services -private finalRegionServerServices services +private finalRegionServerServices services @@ -347,7 +347,7 @@ implements RegionEnvironment -publicRegionEnvironment(RegionCoprocessorimpl, +publicRegionEnvironment(RegionCoprocessorimpl, intpriority, intseq, org.apache.hadoop.conf.Configurationconf, @@ -376,7 +376,7 @@ implements getRegion -publicRegiongetRegion() +publicRegiongetRegion() Specified by: getRegionin interfaceRegionCoprocessorEnvironment @@ -391,7 +391,7 @@ implements getOnlineRegions -publicOnlineRegionsgetOnlineRegions() +publicOnlineRegionsgetOnlineRegions() Specified by: getOnlineRegionsin interfaceRegionCoprocessorEnvironment @@ -406,7 +406,7 @@ implements getConnection -publicConnectiongetConnection() +publicConnectiongetConnection() Description copied from interface:RegionCoprocessorEnvironment Returns the hosts' Connection to the Cluster. Do not close! This is a shared connection with the hosting server. Throws https://docs.oracle.com/javase/8/docs/api/java/lang/UnsupportedOperationException.html?is-external=true; title="class or interface in java.lang">UnsupportedOperationException if you try to close @@ -445,7 +445,7 @@ implements createConnection -publicConnectioncreateConnection(org.apache.hadoop.conf.Configurationconf) +publicConnectioncreateConnection(org.apache.hadoop.conf.Configurationconf) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Description copied from interface:RegionCoprocessorEnvironment Creates a cluster connection using the passed Configuration. @@ -481,7 +481,7 @@ implements getServerName -publicServerNamegetServerName() +publicServerNamegetServerName() Specified by: getServerNamein interfaceRegionCoprocessorEnvironment @@ -496,7 +496,7 @@ implements shutdown -publicvoidshutdown() +publicvoidshutdown() Description copied from class:BaseEnvironment Clean up the environment @@ -511,7 +511,7 @@ implements getSharedData -publichttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true; title="class or interface in java.util.concurrent">ConcurrentMaphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">ObjectgetSharedData() +publichttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true; title="class or interface in
[27/51] [partial] hbase-site git commit: Published site at 31da4d0bce69b3a47066a5df675756087ce4dc60.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a754d895/devapidocs/org/apache/hadoop/hbase/master/assignment/GCMergedRegionsProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/GCMergedRegionsProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/GCMergedRegionsProcedure.html index fa987ee..a13341f 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/GCMergedRegionsProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/GCMergedRegionsProcedure.html @@ -317,7 +317,7 @@ extends AbstractStateMachineTableProcedure -acquireLock, checkTableModifiable, getRegionDir, getUser, releaseLock, < a href="../../../../../../org/apache/hadoop/hbase/master/procedure/AbstractStateMachineTableProcedure.html#releaseSyncLatch--">releaseSyncLatch, setUser +acquireLock, checkTableModifiable, getRegionDir, getUser, preflightChecks, releaseLock, releaseSyncLatch, setUser http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a754d895/devapidocs/org/apache/hadoop/hbase/master/assignment/GCRegionProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/GCRegionProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/GCRegionProcedure.html index 72ff612..671b283 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/GCRegionProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/GCRegionProcedure.html @@ -305,7 +305,7 @@ extends AbstractStateMachineTableProcedure -getRegionDir, getUser, releaseSyncLatch, setUser +getRegionDir, getUser, preflightChecks, releaseSyncLatch, setUser http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a754d895/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html index 347cea7..2881954 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html @@ -129,11 +129,13 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public class MergeTableRegionsProcedure +public class MergeTableRegionsProcedure extends AbstractStateMachineTableProcedureorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.MergeTableRegionsState The procedure to Merge a region in a table. This procedure takes an exclusive table lock since it is working over multiple regions. - It holds the lock for the life of the procedure. + It holds the lock for the life of the procedure. + Throws exception on construction if determines context hostile to merge (cluster going + down or master is shutting down or table is disabled). @@ -520,7 +522,7 @@ extends AbstractStateMachineTableProcedure -checkTableModifiable, getRegionDir, getUser, releaseSyncLatch, setUser +checkTableModifiable, getRegionDir, getUser, preflightChecks, releaseSyncLatch, setUser @@ -563,7 +565,7 @@ extends LOG -private static finalorg.slf4j.Logger LOG +private static finalorg.slf4j.Logger LOG @@ -572,7 +574,7 @@ extends traceEnabled -privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean traceEnabled +privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean traceEnabled @@ -581,7 +583,7 @@ extends lock -private volatileboolean lock +private volatileboolean lock @@ -590,7 +592,7 @@ extends regionLocation -privateServerName regionLocation +privateServerName regionLocation @@ -599,7 +601,7 @@ extends regionsToMerge -privateRegionInfo[] regionsToMerge +privateRegionInfo[] regionsToMerge @@ -608,7 +610,7 @@ extends mergedRegion -privateRegionInfo mergedRegion +privateRegionInfo mergedRegion @@ -617,7 +619,7 @@ extends forcible -privateboolean forcible +privateboolean forcible @@ -634,7 +636,7 @@ extends MergeTableRegionsProcedure -publicMergeTableRegionsProcedure() +publicMergeTableRegionsProcedure() @@ -643,7 +645,7 @@ extends MergeTableRegionsProcedure -publicMergeTableRegionsProcedure(MasterProcedureEnvenv, +publicMergeTableRegionsProcedure(MasterProcedureEnvenv, RegionInforegionToMergeA, RegionInforegionToMergeB) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in
[27/51] [partial] hbase-site git commit: Published site at 6b77786dfc46d25ac5bb5f1c8a4a9eb47b52a604.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html index 71b940f..7ac76dd 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncConnectionImpl.html @@ -106,11 +106,11 @@ private AsyncConnectionImpl -RawAsyncTableImpl.conn +AsyncClientScanner.conn private AsyncConnectionImpl -AsyncBatchRpcRetryingCaller.conn +AsyncRpcRetryingCallerFactory.conn private AsyncConnectionImpl @@ -118,19 +118,19 @@ private AsyncConnectionImpl -RegionCoprocessorRpcChannelImpl.conn +RawAsyncTableImpl.conn -protected AsyncConnectionImpl -AsyncRpcRetryingCaller.conn +private AsyncConnectionImpl +RegionCoprocessorRpcChannelImpl.conn private AsyncConnectionImpl -AsyncClientScanner.conn +AsyncBatchRpcRetryingCaller.conn -private AsyncConnectionImpl -AsyncRpcRetryingCallerFactory.conn +protected AsyncConnectionImpl +AsyncRpcRetryingCaller.conn private AsyncConnectionImpl http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html index e71ca45..d6b1759 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncMasterRequestRpcRetryingCaller.Callable.html @@ -105,13 +105,13 @@ -private AsyncMasterRequestRpcRetryingCaller.CallableT -AsyncMasterRequestRpcRetryingCaller.callable - - private AsyncMasterRequestRpcRetryingCaller.CallableT AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder.callable + +private AsyncMasterRequestRpcRetryingCaller.CallableT +AsyncMasterRequestRpcRetryingCaller.callable + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html index 5f86673..f2374f5 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncProcess.html @@ -106,11 +106,11 @@ private AsyncProcess -BufferedMutatorImpl.ap +HTableMultiplexer.FlushWorker.ap private AsyncProcess -HTableMultiplexer.FlushWorker.ap +BufferedMutatorImpl.ap private AsyncProcess @@ -137,11 +137,11 @@ AsyncProcess -ClusterConnection.getAsyncProcess() +ConnectionImplementation.getAsyncProcess() AsyncProcess -ConnectionImplementation.getAsyncProcess() +ClusterConnection.getAsyncProcess() (package private) AsyncProcess http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegionLocator.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegionLocator.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegionLocator.html index c610e19..9a8d746 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegionLocator.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegionLocator.html @@ -106,11 +106,11 @@ private AsyncRegionLocator -AsyncConnectionImpl.locator +AsyncTableRegionLocatorImpl.locator private AsyncRegionLocator -AsyncTableRegionLocatorImpl.locator +AsyncConnectionImpl.locator http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegistry.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegistry.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegistry.html index 447fe97..e4ecf34 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegistry.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncRegistry.html @@ -126,13 +126,13 @@ -(package private) AsyncRegistry -AsyncConnectionImpl.registry - - private AsyncRegistry AsyncMetaRegionLocator.registry + +(package private) AsyncRegistry +AsyncConnectionImpl.registry +
[27/51] [partial] hbase-site git commit: Published site at 1384da71375427b522b09f06862bb5d629cef52f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d347bde8/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html index d788050..bba78a1 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html @@ -2052,119 +2052,119 @@ service. private TableName -RegionCoprocessorRpcChannel.table +SnapshotDescription.table private TableName -SnapshotDescription.table +RegionCoprocessorRpcChannel.table private TableName -HRegionLocator.tableName +RawAsyncTableImpl.tableName private TableName -ScannerCallableWithReplicas.tableName +RegionServerCallable.tableName protected TableName -ClientScanner.tableName +RegionAdminServiceCallable.tableName private TableName -AsyncClientScanner.tableName +BufferedMutatorImpl.tableName private TableName -AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder.tableName +AsyncProcessTask.tableName private TableName -AsyncRpcRetryingCallerFactory.BatchCallerBuilder.tableName +AsyncProcessTask.Builder.tableName private TableName -RegionInfoBuilder.tableName +AsyncRequestFutureImpl.tableName -private TableName -RegionInfoBuilder.MutableRegionInfo.tableName +protected TableName +TableBuilderBase.tableName private TableName -RawAsyncTableImpl.tableName +AsyncBatchRpcRetryingCaller.tableName private TableName -RegionCoprocessorRpcChannelImpl.tableName +RegionInfoBuilder.tableName private TableName -AsyncTableRegionLocatorImpl.tableName +RegionInfoBuilder.MutableRegionInfo.tableName -protected TableName -RegionAdminServiceCallable.tableName +private TableName +HTable.tableName private TableName -HTable.tableName +TableState.tableName -private TableName -BufferedMutatorImpl.tableName +protected TableName +RpcRetryingCallerWithReadReplicas.tableName -private TableName -AsyncBatchRpcRetryingCaller.tableName +protected TableName +AsyncTableBuilderBase.tableName private TableName -BufferedMutatorParams.tableName +AsyncSingleRequestRpcRetryingCaller.tableName private TableName -HBaseAdmin.TableFuture.tableName +ScannerCallableWithReplicas.tableName -private TableName -AsyncRequestFutureImpl.tableName +protected TableName +RawAsyncHBaseAdmin.TableProcedureBiConsumer.tableName private TableName -AsyncProcessTask.tableName +AsyncTableRegionLocatorImpl.tableName private TableName -AsyncProcessTask.Builder.tableName +HBaseAdmin.TableFuture.tableName -protected TableName -RawAsyncHBaseAdmin.TableProcedureBiConsumer.tableName +private TableName +RegionCoprocessorRpcChannelImpl.tableName -private TableName -RegionServerCallable.tableName +protected TableName +ClientScanner.tableName private TableName -AsyncSingleRequestRpcRetryingCaller.tableName +BufferedMutatorParams.tableName -protected TableName -TableBuilderBase.tableName +private TableName +AsyncClientScanner.tableName -protected TableName -RpcRetryingCallerWithReadReplicas.tableName +private TableName +AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder.tableName -protected TableName -AsyncTableBuilderBase.tableName +private TableName +AsyncRpcRetryingCallerFactory.BatchCallerBuilder.tableName private TableName -TableState.tableName +HRegionLocator.tableName @@ -2206,83 +2206,83 @@ service. TableName -AsyncTable.getName() -Gets the fully qualified table name instance of this table. - +RawAsyncTableImpl.getName() TableName -Table.getName() +RegionLocator.getName() Gets the fully qualified table name instance of this table. TableName -HRegionLocator.getName() +BufferedMutatorImpl.getName() TableName -AsyncTableRegionLocator.getName() -Gets the fully qualified table name instance of the table whose region we want to locate. +BufferedMutator.getName() +Gets the fully qualified table name instance of the table that this BufferedMutator writes to. TableName -AsyncTableImpl.getName() +HTable.getName() TableName -RawAsyncTableImpl.getName() +AsyncBufferedMutator.getName() +Gets the fully qualified table name instance of the table that this + AsyncBufferedMutator writes to. + TableName -AsyncTableRegionLocatorImpl.getName() +Table.getName() +Gets the fully qualified table name instance of this table. + TableName -BufferedMutator.getName() -Gets the fully qualified table name instance of the table that this BufferedMutator writes to. - +AsyncTableImpl.getName() TableName -RegionLocator.getName() +AsyncTable.getName() Gets the fully qualified table name instance of this table. TableName -AsyncBufferedMutatorImpl.getName() +AsyncTableRegionLocatorImpl.getName() TableName -HTable.getName() +AsyncTableRegionLocator.getName() +Gets the fully qualified table
[27/51] [partial] hbase-site git commit: Published site at b7b86839250bf9b295ebc1948826f43a88736d6c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6b94a2f2/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.html b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.html index 2b5efbb..f0821d4 100644 --- a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.html +++ b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceWALReader.html @@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab"; PrevClass -NextClass +NextClass Frames @@ -74,7 +74,7 @@ var activeTableTab = "activeTableTab"; Summary: -Nested| +Nested| Field| Constr| Method @@ -126,7 +126,8 @@ var activeTableTab = "activeTableTab"; @InterfaceStability.Evolving public class ReplicationSourceWALReader extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true; title="class or interface in java.lang">Thread -Reads and filters WAL entries, groups the filtered entries into batches, and puts the batches onto a queue +Reads and filters WAL entries, groups the filtered entries into batches, and puts the batches + onto a queue @@ -139,19 +140,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html Nested Class Summary - -Nested Classes - -Modifier and Type -Class and Description - - -(package private) static class -ReplicationSourceWALReader.WALEntryBatch -Holds a batch of WAL entries to replicate, along with some statistics - - - @@ -182,7 +170,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html currentPosition -protected https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/BlockingQueue.html?is-external=true; title="class or interface in java.util.concurrent">BlockingQueueReplicationSourceWALReader.WALEntryBatch +protected https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/BlockingQueue.html?is-external=true; title="class or interface in java.util.concurrent">BlockingQueueWALEntryBatch entryBatchQueue @@ -222,18 +210,22 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html replicationBatchSizeCapacity +private SerialReplicationChecker +serialReplicationChecker + + private long sleepForRetries - + private ReplicationSource source - + private long totalBufferQuota - + private https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicLong.html?is-external=true; title="class or interface in java.util.concurrent.atomic">AtomicLong totalBufferUsed @@ -317,8 +309,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html protected void -handleEmptyWALEntryBatch(ReplicationSourceWALReader.WALEntryBatchbatch, - org.apache.hadoop.fs.PathcurrentPath) +handleEmptyWALEntryBatch(org.apache.hadoop.fs.PathcurrentPath) private void @@ -329,7 +320,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html isReaderRunning() -private ReplicationSourceWALReader.WALEntryBatch +private WALEntryBatch readWALEntries(WALEntryStreamentryStream) @@ -341,7 +332,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html setReaderRunning(booleanreaderRunning) -ReplicationSourceWALReader.WALEntryBatch +WALEntryBatch take() Retrieves the next batch of WAL entries from the queue, waiting up to the specified time for a batch to become available @@ -349,9 +340,8 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html private void -updateBatchStats(ReplicationSourceWALReader.WALEntryBatchbatch, +updateBatchStats(WALEntryBatchbatch, WAL.Entryentry, -longentryPosition, longentrySize) @@ -443,7 +433,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html entryBatchQueue -protected finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/BlockingQueue.html?is-external=true; title="class or interface in java.util.concurrent">BlockingQueueReplicationSourceWALReader.WALEntryBatch entryBatchQueue +protected finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/BlockingQueue.html?is-external=true; title="class or interface in java.util.concurrent">BlockingQueueWALEntryBatch entryBatchQueue @@ -521,12 +511,21 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html - + totalBufferQuota privatelong totalBufferQuota + + + + + +serialReplicationChecker +private finalSerialReplicationChecker serialReplicationChecker + + @@ -541,7 +540,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html ReplicationSourceWALReader
[27/51] [partial] hbase-site git commit: Published site at 1d25b60831b8cc8f7ad5fd366f1867de5c20d2f3.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb05e3e3/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html -- diff --git a/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html b/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html index 6ea4d25..1900827 100644 --- a/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html +++ b/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html @@ -117,7 +117,7 @@ -static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName +static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName ServerName.EMPTY_SERVER_LIST @@ -166,7 +166,7 @@ static ServerName -ServerName.parseServerName(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringstr) +ServerName.parseServerName(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringstr) static ServerName @@ -177,13 +177,13 @@ static ServerName -ServerName.valueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringserverName) +ServerName.valueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringserverName) Retrieve an instance of ServerName. static ServerName -ServerName.valueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringhostname, +ServerName.valueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringhostname, intport, longstartcode) Retrieve an instance of ServerName. @@ -191,7 +191,7 @@ static ServerName -ServerName.valueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringhostAndPort, +ServerName.valueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringhostAndPort, longstartCode) Retrieve an instance of ServerName. @@ -206,17 +206,17 @@ -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName +https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName ClusterMetrics.getBackupMasterNames() -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName +https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName ClusterStatus.getBackupMasterNames() Deprecated. -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName +https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName ClusterStatus.getBackupMasters() Deprecated. As of release 2.0.0, this will be removed in HBase 3.0.0 @@ -225,27 +225,27 @@ -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName +https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName ClusterMetrics.getDeadServerNames() -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName +https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerName ClusterStatus.getDeadServerNames() Deprecated. -http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,ServerMetrics +https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,ServerMetrics ClusterMetrics.getLiveServerMetrics() -http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,ServerMetrics +https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapServerName,ServerMetrics ClusterStatus.getLiveServerMetrics() Deprecated. -http://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true; title="class or interface
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ec8bf761/devapidocs/org/apache/hadoop/hbase/master/HMaster.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html index b419e6d..8e88e74 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html +++ b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html @@ -2585,7 +2585,7 @@ implements stopProcedureExecutor -privatevoidstopProcedureExecutor() +privatevoidstopProcedureExecutor() @@ -2594,7 +2594,7 @@ implements stopChores -privatevoidstopChores() +privatevoidstopChores() @@ -2603,7 +2603,7 @@ implements getRemoteInetAddress -http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true; title="class or interface in java.net">InetAddressgetRemoteInetAddress(intport, +http://docs.oracle.com/javase/8/docs/api/java/net/InetAddress.html?is-external=true; title="class or interface in java.net">InetAddressgetRemoteInetAddress(intport, longserverStartCode) throws http://docs.oracle.com/javase/8/docs/api/java/net/UnknownHostException.html?is-external=true; title="class or interface in java.net">UnknownHostException @@ -2620,7 +2620,7 @@ implements getMaxBalancingTime -privateintgetMaxBalancingTime() +privateintgetMaxBalancingTime() Returns: Maximum time we should run balancer for @@ -2633,7 +2633,7 @@ implements getMaxRegionsInTransition -privateintgetMaxRegionsInTransition() +privateintgetMaxRegionsInTransition() Returns: Maximum number of regions in transition @@ -2646,7 +2646,7 @@ implements balanceThrottling -privatevoidbalanceThrottling(longnextBalanceStartTime, +privatevoidbalanceThrottling(longnextBalanceStartTime, intmaxRegionsInTransition, longcutoffTime) It first sleep to the next balance plan start time. Meanwhile, throttling by the max @@ -2665,7 +2665,7 @@ implements balance -publicbooleanbalance() +publicbooleanbalance() throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Throws: @@ -2679,7 +2679,7 @@ implements balance -publicbooleanbalance(booleanforce) +publicbooleanbalance(booleanforce) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Throws: @@ -2693,7 +2693,7 @@ implements getRegionNormalizer -publicRegionNormalizergetRegionNormalizer() +publicRegionNormalizergetRegionNormalizer() Specified by: getRegionNormalizerin interfaceMasterServices @@ -2708,7 +2708,7 @@ implements normalizeRegions -publicbooleannormalizeRegions() +publicbooleannormalizeRegions() throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Perform normalization of cluster (invoked by RegionNormalizerChore). @@ -2727,7 +2727,7 @@ implements getClientIdAuditPrefix -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetClientIdAuditPrefix() +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetClientIdAuditPrefix() Specified by: getClientIdAuditPrefixin interfaceMasterServices @@ -2742,7 +2742,7 @@ implements setCatalogJanitorEnabled -publicvoidsetCatalogJanitorEnabled(booleanb) +publicvoidsetCatalogJanitorEnabled(booleanb) Switch for the background CatalogJanitor thread. Used for testing. The thread will continue to run. It will just be a noop if disabled. @@ -2758,7 +2758,7 @@ implements mergeRegions -publiclongmergeRegions(RegionInfo[]regionsToMerge, +publiclongmergeRegions(RegionInfo[]regionsToMerge, booleanforcible, longnonceGroup, longnonce) @@ -2786,7 +2786,7 @@ implements splitRegion -publiclongsplitRegion(RegionInforegionInfo, +publiclongsplitRegion(RegionInforegionInfo, byte[]splitRow, longnonceGroup, longnonce) @@ -2814,7 +2814,7 @@ implements move -publicvoidmove(byte[]encodedRegionName, +publicvoidmove(byte[]encodedRegionName, byte[]destServerName) throws HBaseIOException @@ -2829,7 +2829,7 @@ implements createTable -publiclongcreateTable(TableDescriptortableDescriptor, +publiclongcreateTable(TableDescriptortableDescriptor, byte[][]splitKeys, longnonceGroup, longnonce) @@ -2854,7 +2854,7 @@ implements
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/991224b9/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowColCell.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowColCell.html b/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowColCell.html index d143ef8..4583895 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowColCell.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.LastOnRowColCell.html @@ -258,7 +258,7 @@ 250 251@Override 252public long heapSize() { -253 long sum = HEAP_SIZE_OVERHEAD + estimatedHeapSizeOf(cell); +253 long sum = HEAP_SIZE_OVERHEAD + estimatedSizeOfCell(cell); 254 if (this.tags != null) { 255sum += ClassSize.sizeOf(this.tags); 256 } @@ -454,7 +454,7 @@ 446 447@Override 448public long heapSize() { -449 long sum = HEAP_SIZE_OVERHEAD + estimatedHeapSizeOf(cell); +449 long sum = HEAP_SIZE_OVERHEAD + estimatedSizeOfCell(cell); 450 // this.tags is on heap byte[] 451 if (this.tags != null) { 452sum += ClassSize.sizeOf(this.tags); @@ -2791,192 +2791,193 @@ 2783 * {@link HeapSize} we call {@link HeapSize#heapSize()} so cell can give a correct value. In other 2784 * cases we just consider the bytes occupied by the cell components ie. row, CF, qualifier, 2785 * timestamp, type, value and tags. -2786 * @param cell -2787 * @return estimate of the heap space -2788 */ -2789 public static long estimatedHeapSizeOf(final Cell cell) { -2790if (cell instanceof HeapSize) { -2791 return ((HeapSize) cell).heapSize(); -2792} -2793// TODO: Add sizing of references that hold the row, family, etc., arrays. -2794return estimatedSerializedSizeOf(cell); -2795 } -2796 -2797 /** -2798 * This method exists just to encapsulate how we serialize keys. To be replaced by a factory that -2799 * we query to figure what the Cell implementation is and then, what serialization engine to use -2800 * and further, how to serialize the key for inclusion in hfile index. TODO. -2801 * @param cell -2802 * @return The key portion of the Cell serialized in the old-school KeyValue way or null if passed -2803 * a null codecell/code -2804 */ -2805 public static byte[] getCellKeySerializedAsKeyValueKey(final Cell cell) { -2806if (cell == null) return null; -2807byte[] b = new byte[KeyValueUtil.keyLength(cell)]; -2808KeyValueUtil.appendKeyTo(cell, b, 0); -2809return b; -2810 } -2811 -2812 /** -2813 * Create a Cell that is smaller than all other possible Cells for the given Cell's row. -2814 * @param cell -2815 * @return First possible Cell on passed Cell's row. -2816 */ -2817 public static Cell createFirstOnRow(final Cell cell) { -2818if (cell instanceof ByteBufferExtendedCell) { -2819 return new FirstOnRowByteBufferExtendedCell( -2820 ((ByteBufferExtendedCell) cell).getRowByteBuffer(), -2821 ((ByteBufferExtendedCell) cell).getRowPosition(), cell.getRowLength()); -2822} -2823return new FirstOnRowCell(cell.getRowArray(), cell.getRowOffset(), cell.getRowLength()); -2824 } -2825 -2826 public static Cell createFirstOnRow(final byte[] row, int roffset, short rlength) { -2827return new FirstOnRowCell(row, roffset, rlength); -2828 } -2829 -2830 public static Cell createFirstOnRow(final byte[] row, final byte[] family, final byte[] col) { -2831return createFirstOnRow(row, 0, (short) row.length, family, 0, (byte) family.length, col, 0, -2832col.length); -2833 } -2834 -2835 public static Cell createFirstOnRow(final byte[] row, int roffset, short rlength, -2836 final byte[] family, int foffset, byte flength, final byte[] col, int coffset, int clength) { -2837return new FirstOnRowColCell(row, roffset, rlength, family, foffset, flength, col, coffset, -2838clength); -2839 } -2840 -2841 public static Cell createFirstOnRow(final byte[] row) { -2842return createFirstOnRow(row, 0, (short) row.length); -2843 } -2844 -2845 public static Cell createFirstOnRowFamily(Cell cell, byte[] fArray, int foff, int flen) { -2846if (cell instanceof ByteBufferExtendedCell) { -2847 return new FirstOnRowColByteBufferExtendedCell( -2848 ((ByteBufferExtendedCell) cell).getRowByteBuffer(), -2849 ((ByteBufferExtendedCell) cell).getRowPosition(), cell.getRowLength(), -2850 ByteBuffer.wrap(fArray), foff, (byte) flen, HConstants.EMPTY_BYTE_BUFFER, 0, 0); -2851} -2852return new FirstOnRowColCell(cell.getRowArray(), cell.getRowOffset(), cell.getRowLength(), -2853fArray, foff, (byte) flen, HConstants.EMPTY_BYTE_ARRAY, 0, 0); -2854 } -2855 -2856 public static Cell createFirstOnRowCol(final Cell cell) { -2857if (cell
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html index cda7c78..93f650f 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceState.html @@ -269,7 +269,7 @@ 261this.rpcControllerFactory = RpcControllerFactory.instantiate(conf); 262this.rpcCallerFactory = RpcRetryingCallerFactory.instantiate(conf, interceptor, this.stats); 263this.backoffPolicy = ClientBackoffPolicyFactory.create(conf); -264this.asyncProcess = new AsyncProcess(this, conf, rpcCallerFactory, false, rpcControllerFactory); +264this.asyncProcess = new AsyncProcess(this, conf, rpcCallerFactory, rpcControllerFactory); 265if (conf.getBoolean(CLIENT_SIDE_METRICS_ENABLED_KEY, false)) { 266 this.metrics = new MetricsConnection(this); 267} else { http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html index cda7c78..93f650f 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.MasterServiceStubMaker.html @@ -269,7 +269,7 @@ 261this.rpcControllerFactory = RpcControllerFactory.instantiate(conf); 262this.rpcCallerFactory = RpcRetryingCallerFactory.instantiate(conf, interceptor, this.stats); 263this.backoffPolicy = ClientBackoffPolicyFactory.create(conf); -264this.asyncProcess = new AsyncProcess(this, conf, rpcCallerFactory, false, rpcControllerFactory); +264this.asyncProcess = new AsyncProcess(this, conf, rpcCallerFactory, rpcControllerFactory); 265if (conf.getBoolean(CLIENT_SIDE_METRICS_ENABLED_KEY, false)) { 266 this.metrics = new MetricsConnection(this); 267} else { http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html index cda7c78..93f650f 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html @@ -269,7 +269,7 @@ 261this.rpcControllerFactory = RpcControllerFactory.instantiate(conf); 262this.rpcCallerFactory = RpcRetryingCallerFactory.instantiate(conf, interceptor, this.stats); 263this.backoffPolicy = ClientBackoffPolicyFactory.create(conf); -264this.asyncProcess = new AsyncProcess(this, conf, rpcCallerFactory, false, rpcControllerFactory); +264this.asyncProcess = new AsyncProcess(this, conf, rpcCallerFactory, rpcControllerFactory); 265if (conf.getBoolean(CLIENT_SIDE_METRICS_ENABLED_KEY, false)) { 266 this.metrics = new MetricsConnection(this); 267} else { http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.html index cda7c78..93f650f 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.html @@ -269,7 +269,7 @@ 261this.rpcControllerFactory = RpcControllerFactory.instantiate(conf); 262this.rpcCallerFactory = RpcRetryingCallerFactory.instantiate(conf, interceptor, this.stats); 263this.backoffPolicy =
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/94208cfe/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html index 90f52b0..2ac1b78 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocateType.html @@ -106,7 +106,7 @@ private RegionLocateType -AsyncSingleRequestRpcRetryingCaller.locateType +AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder.locateType RegionLocateType @@ -114,7 +114,7 @@ private RegionLocateType -AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder.locateType +AsyncSingleRequestRpcRetryingCaller.locateType http://git-wip-us.apache.org/repos/asf/hbase-site/blob/94208cfe/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html index e062eb5..fbe0658 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionLocator.html @@ -230,14 +230,14 @@ service. private RegionLocator -HFileOutputFormat2.TableInfo.regionLocator - - -private RegionLocator TableInputFormatBase.regionLocator The RegionLocator of the table. + +private RegionLocator +HFileOutputFormat2.TableInfo.regionLocator + @@ -248,15 +248,15 @@ service. -RegionLocator -HFileOutputFormat2.TableInfo.getRegionLocator() - - protected RegionLocator TableInputFormatBase.getRegionLocator() Allows subclasses to get the RegionLocator. + +RegionLocator +HFileOutputFormat2.TableInfo.getRegionLocator() +
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cd17dc5/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html index ded834e..9c4acf5 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html @@ -330,8 +330,10 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncHBaseAdmin.addColumnFamily(TableNametableName, - ColumnFamilyDescriptorcolumnFamily) +AsyncAdmin.addColumnFamily(TableNametableName, + ColumnFamilyDescriptorcolumnFamily) +Add a column family to an existing table. + void @@ -341,20 +343,18 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -void -HBaseAdmin.addColumnFamily(TableNametableName, +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void +RawAsyncHBaseAdmin.addColumnFamily(TableNametableName, ColumnFamilyDescriptorcolumnFamily) -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncAdmin.addColumnFamily(TableNametableName, - ColumnFamilyDescriptorcolumnFamily) -Add a column family to an existing table. - +void +HBaseAdmin.addColumnFamily(TableNametableName, + ColumnFamilyDescriptorcolumnFamily) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -RawAsyncHBaseAdmin.addColumnFamily(TableNametableName, +AsyncHBaseAdmin.addColumnFamily(TableNametableName, ColumnFamilyDescriptorcolumnFamily) @@ -396,8 +396,10 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncHBaseAdmin.modifyColumnFamily(TableNametableName, - ColumnFamilyDescriptorcolumnFamily) +AsyncAdmin.modifyColumnFamily(TableNametableName, + ColumnFamilyDescriptorcolumnFamily) +Modify an existing column family on a table. + void @@ -407,20 +409,18 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -void -HBaseAdmin.modifyColumnFamily(TableNametableName, +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void +RawAsyncHBaseAdmin.modifyColumnFamily(TableNametableName, ColumnFamilyDescriptorcolumnFamily) -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncAdmin.modifyColumnFamily(TableNametableName, - ColumnFamilyDescriptorcolumnFamily) -Modify an existing column family on a table. - +void +HBaseAdmin.modifyColumnFamily(TableNametableName, + ColumnFamilyDescriptorcolumnFamily) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html index 53d69df..597d1ba 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html @@ -355,24 +355,24 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. void -BufferedMutatorImpl.mutate(Mutationm) - - -void BufferedMutator.mutate(Mutationmutation) Sends a Mutation to the table. + +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void +AsyncBufferedMutatorImpl.mutate(Mutationmutation) + +void +BufferedMutatorImpl.mutate(Mutationm) + + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void AsyncBufferedMutator.mutate(Mutationmutation) Sends a Mutation to the table. - -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncBufferedMutatorImpl.mutate(Mutationmutation) - @@ -390,24 +390,24 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. void -BufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends Mutationms) - - -void BufferedMutator.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends Mutationmutations) Send some Mutations to the table. + +http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void +AsyncBufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends Mutationmutations) + +void +BufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends Mutationms) + + http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void AsyncBufferedMutator.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends Mutationmutations) Send some Mutations to the table. - -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncBufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends Mutationmutations) - static RowMutations RowMutations.of(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends Mutationmutations) @@ -543,15 +543,15 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. MutationSerialization.getDeserializer(http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">ClassMutationc)
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html index 1d50582..7eb7661 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotDisabledException.html @@ -104,14 +104,14 @@ void -HMaster.checkTableModifiable(TableNametableName) - - -void MasterServices.checkTableModifiable(TableNametableName) Check table is modifiable; i.e. + +void +HMaster.checkTableModifiable(TableNametableName) + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html index 5eebc7c..b1a475b 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableNotFoundException.html @@ -144,6 +144,19 @@ Uses of TableNotFoundException in org.apache.hadoop.hbase.master + +Subclasses of TableNotFoundException in org.apache.hadoop.hbase.master + +Modifier and Type +Class and Description + + + +static class +TableStateManager.TableStateNotFoundException + + + Methods in org.apache.hadoop.hbase.master that throw TableNotFoundException @@ -157,14 +170,14 @@ void -HMaster.checkTableModifiable(TableNametableName) - - -void MasterServices.checkTableModifiable(TableNametableName) Check table is modifiable; i.e. + +void +HMaster.checkTableModifiable(TableNametableName) + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html b/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html index 594e74b..90f6967 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/Tag.html @@ -243,6 +243,10 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTag +PrivateCellUtil.getTags(Cellcell) + + +static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTag CellUtil.getTags(Cellcell) Deprecated. As of release 2.0.0, this will be removed in HBase 3.0.0. @@ -250,10 +254,6 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. - -static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTag -PrivateCellUtil.getTags(Cellcell) - static http://docs.oracle.com/javase/8/docs/api/java/util/Iterator.html?is-external=true; title="class or interface in java.util">IteratorTag CellUtil.tagsIterator(byte[]tags, @@ -395,6 +395,11 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. static Cell +PrivateCellUtil.createCell(Cellcell, + http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) + + +static Cell CellUtil.createCell(Cellcell, http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) Deprecated. @@ -402,11 +407,6 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. - -static Cell -PrivateCellUtil.createCell(Cellcell, - http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) - static byte[] TagUtil.fromList(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) @@ -415,16 +415,16 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -RawCellBuilder -RawCellBuilder.setTags(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) +ExtendedCellBuilder +ExtendedCellBuilderImpl.setTags(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTagtags) -ExtendedCellBuilder
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html index 53d69df..597d1ba 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Mutation.html @@ -355,24 +355,24 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. void -BufferedMutatorImpl.mutate(Mutationm) - - -void BufferedMutator.mutate(Mutationmutation) Sends a Mutation to the table. + +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void +AsyncBufferedMutatorImpl.mutate(Mutationmutation) + +void +BufferedMutatorImpl.mutate(Mutationm) + + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void AsyncBufferedMutator.mutate(Mutationmutation) Sends a Mutation to the table. - -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncBufferedMutatorImpl.mutate(Mutationmutation) - @@ -390,24 +390,24 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. void -BufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends Mutationms) - - -void BufferedMutator.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends Mutationmutations) Send some Mutations to the table. + +http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void +AsyncBufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends Mutationmutations) + +void +BufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends Mutationms) + + http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void AsyncBufferedMutator.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends Mutationmutations) Send some Mutations to the table. - -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncBufferedMutatorImpl.mutate(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends Mutationmutations) - static RowMutations RowMutations.of(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends Mutationmutations) @@ -543,15 +543,15 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. MutationSerialization.getDeserializer(http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">ClassMutationc)
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0ab8335e/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html index 232a8b4..aeaf9fe 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html @@ -495,7 +495,7 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. private static HRegionLocation -AsyncMetaTableAccessor.getRegionLocation(Resultr, +MetaTableAccessor.getRegionLocation(Resultr, RegionInforegionInfo, intreplicaId) Returns the HRegionLocation parsed from the given meta row Result @@ -504,7 +504,7 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. private static HRegionLocation -MetaTableAccessor.getRegionLocation(Resultr, +AsyncMetaTableAccessor.getRegionLocation(Resultr, RegionInforegionInfo, intreplicaId) Returns the HRegionLocation parsed from the given meta row Result @@ -944,7 +944,9 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -AsyncHBaseAdmin.getRegions(ServerNameserverName) +AsyncAdmin.getRegions(ServerNameserverName) +Get all the online regions on a region server. + http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo @@ -953,22 +955,22 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -HBaseAdmin.getRegions(ServerNamesn) +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo +RawAsyncHBaseAdmin.getRegions(ServerNameserverName) -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -AsyncAdmin.getRegions(ServerNameserverName) -Get all the online regions on a region server. - +http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo +HBaseAdmin.getRegions(ServerNamesn) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -RawAsyncHBaseAdmin.getRegions(ServerNameserverName) +AsyncHBaseAdmin.getRegions(ServerNameserverName) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -AsyncHBaseAdmin.getRegions(TableNametableName) +AsyncAdmin.getRegions(TableNametableName) +Get the regions of a given table. + http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo @@ -977,18 +979,16 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -HBaseAdmin.getRegions(TableNametableName) +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo +RawAsyncHBaseAdmin.getRegions(TableNametableName) -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html index ad601c4..53e455f 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html @@ -1117,1183 +1117,1186 @@ 1109 @Nullable 1110 public static TableState getTableState(Connection conn, TableName tableName) throws IOException { -1112Table metaHTable = getMetaHTable(conn); -1113Get get = new Get(tableName.getName()).addColumn(getTableFamily(), getTableStateColumn()); -1114long time = EnvironmentEdgeManager.currentTime(); -1115get.setTimeRange(0, time); -1116Result result = -1117metaHTable.get(get); -1118return getTableState(result); -1119 } -1120 -1121 /** -1122 * Fetch table states from META table -1123 * @param conn connection to use -1124 * @return map {tableName -gt; state} -1125 * @throws IOException -1126 */ -1127 public static MapTableName, TableState getTableStates(Connection conn) -1128 throws IOException { -1129final MapTableName, TableState states = new LinkedHashMap(); -1130Visitor collector = new Visitor() { -1131 @Override -1132 public boolean visit(Result r) throws IOException { -1133TableState state = getTableState(r); -1134if (state != null) -1135 states.put(state.getTableName(), state); -1136return true; -1137 } -1138}; -1139fullScanTables(conn, collector); -1140return states; -1141 } -1142 -1143 /** -1144 * Updates state in META -1145 * @param conn connection to use -1146 * @param tableName table to look for -1147 * @throws IOException -1148 */ -1149 public static void updateTableState(Connection conn, TableName tableName, -1150 TableState.State actual) throws IOException { -1151updateTableState(conn, new TableState(tableName, actual)); -1152 } -1153 -1154 /** -1155 * Decode table state from META Result. -1156 * Should contain cell from HConstants.TABLE_FAMILY -1157 * @param r result -1158 * @return null if not found -1159 * @throws IOException -1160 */ -1161 @Nullable -1162 public static TableState getTableState(Result r) -1163 throws IOException { -1164Cell cell = r.getColumnLatestCell(getTableFamily(), getTableStateColumn()); -1165if (cell == null) return null; -1166try { -1167 return TableState.parseFrom(TableName.valueOf(r.getRow()), -1168 Arrays.copyOfRange(cell.getValueArray(), -1169 cell.getValueOffset(), cell.getValueOffset() + cell.getValueLength())); -1170} catch (DeserializationException e) { -1171 throw new IOException(e); -1172} -1173 -1174 } -1175 -1176 /** -1177 * Implementations 'visit' a catalog table row. -1178 */ -1179 public interface Visitor { -1180/** -1181 * Visit the catalog table row. -1182 * @param r A row from catalog table -1183 * @return True if we are to proceed scanning the table, else false if -1184 * we are to stop now. -1185 */ -1186boolean visit(final Result r) throws IOException; -1187 } -1188 -1189 /** -1190 * Implementations 'visit' a catalog table row but with close() at the end. -1191 */ -1192 public interface CloseableVisitor extends Visitor, Closeable { -1193 } -1194 -1195 /** -1196 * A {@link Visitor} that collects content out of passed {@link Result}. -1197 */ -1198 static abstract class CollectingVisitorT implements Visitor { -1199final ListT results = new ArrayList(); -1200@Override -1201public boolean visit(Result r) throws IOException { -1202 if (r == null || r.isEmpty()) return true; -1203 add(r); -1204 return true; -1205} -1206 -1207abstract void add(Result r); -1208 -1209/** -1210 * @return Collected results; wait till visits complete to collect all -1211 * possible results -1212 */ -1213ListT getResults() { -1214 return this.results; -1215} -1216 } -1217 -1218 /** -1219 * Collects all returned. -1220 */ -1221 static class CollectAllVisitor extends CollectingVisitorResult { -1222@Override -1223void add(Result r) { -1224 this.results.add(r); -1225} -1226 } -1227 -1228 /** -1229 * A Visitor that skips offline regions and split parents -1230 */ -1231 public static abstract class DefaultVisitorBase implements Visitor { -1232 -1233public DefaultVisitorBase() { -1234 super(); -1235} -1236 -1237public abstract boolean visitInternal(Result rowResult) throws IOException; -1238 -1239
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1f2eeb22/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html b/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html index ea234c2..34c375d 100644 --- a/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html +++ b/devapidocs/org/apache/hadoop/hbase/util/CommonFSUtils.html @@ -567,7 +567,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? warningMap -private static finalhttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maporg.apache.hadoop.fs.FileSystem,http://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean warningMap +private static finalhttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maporg.apache.hadoop.fs.FileSystem,http://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean warningMap @@ -972,7 +972,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? isValidWALRootDir -private staticbooleanisValidWALRootDir(org.apache.hadoop.fs.PathwalDir, +private staticbooleanisValidWALRootDir(org.apache.hadoop.fs.PathwalDir, org.apache.hadoop.conf.Configurationc) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException @@ -987,7 +987,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? getTableDir -public staticorg.apache.hadoop.fs.PathgetTableDir(org.apache.hadoop.fs.Pathrootdir, +public staticorg.apache.hadoop.fs.PathgetTableDir(org.apache.hadoop.fs.Pathrootdir, TableNametableName) Returns the Path object representing the table directory under path rootdir @@ -1006,7 +1006,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? getTableName -public staticTableNamegetTableName(org.apache.hadoop.fs.PathtablePath) +public staticTableNamegetTableName(org.apache.hadoop.fs.PathtablePath) Returns the TableName object representing the table directory under path rootdir @@ -1024,7 +1024,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? getNamespaceDir -public staticorg.apache.hadoop.fs.PathgetNamespaceDir(org.apache.hadoop.fs.Pathrootdir, +public staticorg.apache.hadoop.fs.PathgetNamespaceDir(org.apache.hadoop.fs.Pathrootdir, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringnamespace) Returns the Path object representing the namespace directory under path rootdir @@ -1043,7 +1043,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? setStoragePolicy -public staticvoidsetStoragePolicy(org.apache.hadoop.fs.FileSystemfs, +public staticvoidsetStoragePolicy(org.apache.hadoop.fs.FileSystemfs, org.apache.hadoop.conf.Configurationconf, org.apache.hadoop.fs.Pathpath, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringpolicyKey, @@ -1074,7 +1074,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? setStoragePolicy -public staticvoidsetStoragePolicy(org.apache.hadoop.fs.FileSystemfs, +public staticvoidsetStoragePolicy(org.apache.hadoop.fs.FileSystemfs, org.apache.hadoop.fs.Pathpath, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringstoragePolicy) Sets storage policy for given path. @@ -1101,7 +1101,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? invokeSetStoragePolicy -private staticvoidinvokeSetStoragePolicy(org.apache.hadoop.fs.FileSystemfs, +private staticvoidinvokeSetStoragePolicy(org.apache.hadoop.fs.FileSystemfs, org.apache.hadoop.fs.Pathpath, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringstoragePolicy) @@ -1112,7 +1112,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? isHDFS -public staticbooleanisHDFS(org.apache.hadoop.conf.Configurationconf) +public
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.html b/testdevapidocs/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.html index 633812a..d239905 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.html +++ b/testdevapidocs/org/apache/hadoop/hbase/TestServerSideScanMetricsFromClientSide.html @@ -109,7 +109,7 @@ var activeTableTab = "activeTableTab"; -public class TestServerSideScanMetricsFromClientSide +public class TestServerSideScanMetricsFromClientSide extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object @@ -134,64 +134,68 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? CELL_HEAP_SIZE +static HBaseClassTestRule +CLASS_RULE + + private static byte[][] FAMILIES - + private static byte[] FAMILY - + private static int NUM_COLS - + private static int NUM_FAMILIES - + private static int NUM_QUALIFIERS - + private static int NUM_ROWS - + private static byte[] QUALIFIER - + private static byte[][] QUALIFIERS - + private static byte[] ROW - + private static byte[][] ROWS - + private static org.apache.hadoop.hbase.client.Table TABLE - + private static org.apache.hadoop.hbase.TableName TABLE_NAME Table configuration - + private static HBaseTestingUtility TEST_UTIL - + private static byte[] VALUE - + private static int VALUE_SIZE @@ -316,13 +320,22 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? Field Detail + + + + + +CLASS_RULE +public static finalHBaseClassTestRule CLASS_RULE + + TEST_UTIL -private static finalHBaseTestingUtility TEST_UTIL +private static finalHBaseTestingUtility TEST_UTIL @@ -331,7 +344,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? TABLE -private staticorg.apache.hadoop.hbase.client.Table TABLE +private staticorg.apache.hadoop.hbase.client.Table TABLE @@ -340,7 +353,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? TABLE_NAME -private staticorg.apache.hadoop.hbase.TableName TABLE_NAME +private staticorg.apache.hadoop.hbase.TableName TABLE_NAME Table configuration @@ -350,7 +363,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? NUM_ROWS -private staticint NUM_ROWS +private staticint NUM_ROWS @@ -359,7 +372,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? ROW -private staticbyte[] ROW +private staticbyte[] ROW @@ -368,7 +381,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? ROWS -private staticbyte[][] ROWS +private staticbyte[][] ROWS @@ -377,7 +390,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? NUM_FAMILIES -private staticint NUM_FAMILIES +private staticint NUM_FAMILIES @@ -386,7 +399,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? FAMILY -private staticbyte[] FAMILY +private staticbyte[] FAMILY @@ -395,7 +408,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? FAMILIES -private staticbyte[][] FAMILIES +private staticbyte[][] FAMILIES @@ -404,7 +417,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? NUM_QUALIFIERS -private staticint NUM_QUALIFIERS +private staticint NUM_QUALIFIERS @@ -413,7 +426,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? QUALIFIER -private staticbyte[] QUALIFIER +private staticbyte[] QUALIFIER @@ -422,7 +435,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? QUALIFIERS -private staticbyte[][] QUALIFIERS +private staticbyte[][] QUALIFIERS @@ -431,7 +444,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? VALUE_SIZE -private staticint VALUE_SIZE +private staticint VALUE_SIZE @@ -440,7 +453,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? VALUE -private staticbyte[] VALUE +private staticbyte[] VALUE @@ -449,7 +462,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? NUM_COLS -private staticint NUM_COLS +private staticint NUM_COLS @@ -458,7 +471,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? CELL_HEAP_SIZE -private staticlong CELL_HEAP_SIZE +private staticlong CELL_HEAP_SIZE @@ -475,7 +488,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? TestServerSideScanMetricsFromClientSide -publicTestServerSideScanMetricsFromClientSide()
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa7ffc92/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html index eb9e252..667152a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html @@ -28,22 +28,22 @@ 020 021import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH; 022import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH_DESC; -023import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH; -024import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC; -025import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER; -026import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC; -027import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET; -028import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC; -029import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC; -030import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE; -031import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC; -032import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC; -033import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS; -034import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC; -035import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME; -036import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC; -037import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG; -038import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC; +023import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG; +024import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC; +025import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH; +026import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC; +027import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER; +028import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC; +029import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET; +030import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC; +031import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC; +032import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE; +033import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC; +034import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC; +035import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS; +036import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC; +037import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME; +038import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC; 039 040import java.io.IOException; 041import java.net.URI; @@ -70,194 +70,194 @@ 062import org.apache.hadoop.hbase.backup.util.BackupUtils; 063import org.apache.hadoop.hbase.client.Connection; 064import org.apache.hadoop.hbase.client.ConnectionFactory; -065import org.apache.hbase.thirdparty.com.google.common.collect.Lists; -066import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -067import org.apache.yetus.audience.InterfaceAudience; -068 -069/** -070 * General backup commands, options and usage messages -071 */ -072 +065import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; +066import org.apache.yetus.audience.InterfaceAudience; +067 +068import org.apache.hbase.thirdparty.com.google.common.collect.Lists; +069 +070/** +071 * General backup commands, options and usage messages +072 */ 073@InterfaceAudience.Private 074public final class BackupCommands { -075 -076 public final static String INCORRECT_USAGE = "Incorrect usage"; -077 -078 public final static String TOP_LEVEL_NOT_ALLOWED = -079 "Top level (root) folder is not allowed to be a backup destination"; -080 -081 public static final String USAGE = "Usage: hbase
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/96e5e102/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/NoLimitScannerContext.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/NoLimitScannerContext.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/NoLimitScannerContext.html index 0164ad3..fd16dee 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/NoLimitScannerContext.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/NoLimitScannerContext.html @@ -76,64 +76,54 @@ 068 } 069 070 @Override -071 void setTimeProgress(long timeProgress) { +071 void setProgress(int batchProgress, long sizeProgress, long heapSizeProgress) { 072// Do nothing. NoLimitScannerContext instances are immutable post-construction 073 } 074 075 @Override -076 void updateTimeProgress() { +076 void clearProgress() { 077// Do nothing. NoLimitScannerContext instances are immutable post-construction 078 } 079 080 @Override -081 void setProgress(int batchProgress, long sizeProgress, long heapSizeProgress, long timeProgress) { +081 void setSizeLimitScope(LimitScope scope) { 082// Do nothing. NoLimitScannerContext instances are immutable post-construction 083 } 084 085 @Override -086 void clearProgress() { +086 void setTimeLimitScope(LimitScope scope) { 087// Do nothing. NoLimitScannerContext instances are immutable post-construction 088 } 089 090 @Override -091 void setSizeLimitScope(LimitScope scope) { +091 NextState setScannerState(NextState state) { 092// Do nothing. NoLimitScannerContext instances are immutable post-construction -093 } -094 -095 @Override -096 void setTimeLimitScope(LimitScope scope) { -097// Do nothing. NoLimitScannerContext instances are immutable post-construction -098 } -099 -100 @Override -101 NextState setScannerState(NextState state) { -102// Do nothing. NoLimitScannerContext instances are immutable post-construction -103return state; -104 } -105 -106 @Override -107 boolean checkBatchLimit(LimitScope checkerScope) { -108// No limits can be specified, thus return false to indicate no limit has been reached. -109return false; -110 } -111 -112 @Override -113 boolean checkSizeLimit(LimitScope checkerScope) { -114// No limits can be specified, thus return false to indicate no limit has been reached. -115return false; -116 } -117 -118 @Override -119 boolean checkTimeLimit(LimitScope checkerScope) { -120// No limits can be specified, thus return false to indicate no limit has been reached. -121return false; -122 } -123 -124 @Override -125 boolean checkAnyLimitReached(LimitScope checkerScope) { -126return false; -127 } -128} +093return state; +094 } +095 +096 @Override +097 boolean checkBatchLimit(LimitScope checkerScope) { +098// No limits can be specified, thus return false to indicate no limit has been reached. +099return false; +100 } +101 +102 @Override +103 boolean checkSizeLimit(LimitScope checkerScope) { +104// No limits can be specified, thus return false to indicate no limit has been reached. +105return false; +106 } +107 +108 @Override +109 boolean checkTimeLimit(LimitScope checkerScope) { +110// No limits can be specified, thus return false to indicate no limit has been reached. +111return false; +112 } +113 +114 @Override +115 boolean checkAnyLimitReached(LimitScope checkerScope) { +116return false; +117 } +118}
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8118541f/devapidocs/org/apache/hadoop/hbase/io/hfile/LruBlockCache.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/io/hfile/LruBlockCache.html b/devapidocs/org/apache/hadoop/hbase/io/hfile/LruBlockCache.html index f8c49cb..e3762b3 100644 --- a/devapidocs/org/apache/hadoop/hbase/io/hfile/LruBlockCache.html +++ b/devapidocs/org/apache/hadoop/hbase/io/hfile/LruBlockCache.html @@ -1265,7 +1265,7 @@ implements CACHE_FIXED_OVERHEAD -public static finallong CACHE_FIXED_OVERHEAD +public static finallong CACHE_FIXED_OVERHEAD @@ -1434,7 +1434,7 @@ implements cacheBlock -publicvoidcacheBlock(BlockCacheKeycacheKey, +publicvoidcacheBlock(BlockCacheKeycacheKey, Cacheablebuf) Cache the block with the specified name and buffer. @@ -1453,7 +1453,7 @@ implements updateSizeMetrics -privatelongupdateSizeMetrics(LruCachedBlockcb, +privatelongupdateSizeMetrics(LruCachedBlockcb, booleanevict) Helper function that updates the local size counter and also updates any per-cf or per-blocktype metrics it can discern from given @@ -1466,7 +1466,7 @@ implements getBlock -publicCacheablegetBlock(BlockCacheKeycacheKey, +publicCacheablegetBlock(BlockCacheKeycacheKey, booleancaching, booleanrepeat, booleanupdateCacheMetrics) @@ -1492,7 +1492,7 @@ implements containsBlock -publicbooleancontainsBlock(BlockCacheKeycacheKey) +publicbooleancontainsBlock(BlockCacheKeycacheKey) Whether the cache contains block with specified cacheKey Returns: @@ -1506,7 +1506,7 @@ implements evictBlock -publicbooleanevictBlock(BlockCacheKeycacheKey) +publicbooleanevictBlock(BlockCacheKeycacheKey) Description copied from interface:BlockCache Evict block from cache. @@ -1525,7 +1525,7 @@ implements evictBlocksByHfileName -publicintevictBlocksByHfileName(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringhfileName) +publicintevictBlocksByHfileName(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringhfileName) Evicts all blocks for a specific HFile. This is an expensive operation implemented as a linear-time search through all blocks in the cache. Ideally this should be a search in a log-access-time map. @@ -1546,7 +1546,7 @@ implements evictBlock -protectedlongevictBlock(LruCachedBlockblock, +protectedlongevictBlock(LruCachedBlockblock, booleanevictedByEvictionProcess) Evict the block, and it will be cached by the victim handler if exists block may be read again later @@ -1565,7 +1565,7 @@ implements runEviction -privatevoidrunEviction() +privatevoidrunEviction() Multi-threaded call to run the eviction process. @@ -1575,7 +1575,7 @@ implements isEvictionInProgress -booleanisEvictionInProgress() +booleanisEvictionInProgress() @@ -1584,7 +1584,7 @@ implements getOverhead -longgetOverhead() +longgetOverhead() @@ -1593,7 +1593,7 @@ implements evict -voidevict() +voidevict() Eviction method. @@ -1603,7 +1603,7 @@ implements toString -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringtoString() +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringtoString() Overrides: http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--; title="class or interface in java.lang">toStringin classhttp://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object @@ -1616,7 +1616,7 @@ implements getMaxSize -publiclonggetMaxSize() +publiclonggetMaxSize() Get the maximum size of this cache. Specified by: @@ -1632,7 +1632,7 @@ implements getCurrentSize -publiclonggetCurrentSize() +publiclonggetCurrentSize() Description copied from interface:BlockCache Returns the occupied size of the block cache, in bytes. @@ -1649,7 +1649,7 @@ implements getCurrentDataSize -publiclonggetCurrentDataSize() +publiclonggetCurrentDataSize() Description copied from interface:BlockCache Returns the occupied size of data blocks, in bytes. @@ -1666,7 +1666,7 @@ implements getFreeSize -publiclonggetFreeSize() +publiclonggetFreeSize() Description copied from interface:BlockCache Returns the free size of the block cache, in bytes. @@ -1683,7 +1683,7 @@ implements size -publiclongsize() +publiclongsize() Description copied from interface:BlockCache Returns the total size of the block cache, in bytes. @@ -1700,7 +1700,7 @@ implements getBlockCount
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/14db89d7/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.ServerEventsListenerThread.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.ServerEventsListenerThread.html b/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.ServerEventsListenerThread.html index e05510e..2e114d8 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.ServerEventsListenerThread.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.ServerEventsListenerThread.html @@ -54,891 +54,884 @@ 046import org.apache.hadoop.hbase.HTableDescriptor; 047import org.apache.hadoop.hbase.MetaTableAccessor; 048import org.apache.hadoop.hbase.MetaTableAccessor.DefaultVisitorBase; -049import org.apache.hadoop.hbase.NamespaceDescriptor; -050import org.apache.hadoop.hbase.ServerName; -051import org.apache.hadoop.hbase.TableName; -052import org.apache.hadoop.hbase.client.ClusterConnection; -053import org.apache.hadoop.hbase.client.Delete; -054import org.apache.hadoop.hbase.client.Get; -055import org.apache.hadoop.hbase.client.Mutation; -056import org.apache.hadoop.hbase.client.Put; -057import org.apache.hadoop.hbase.client.RegionInfo; -058import org.apache.hadoop.hbase.client.Result; -059import org.apache.hadoop.hbase.client.Scan; -060import org.apache.hadoop.hbase.client.Table; -061import org.apache.hadoop.hbase.client.TableState; -062import org.apache.hadoop.hbase.constraint.ConstraintException; -063import org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint; -064import org.apache.hadoop.hbase.exceptions.DeserializationException; -065import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel; -066import org.apache.hadoop.hbase.master.MasterServices; -067import org.apache.hadoop.hbase.master.ServerListener; -068import org.apache.hadoop.hbase.master.TableStateManager; -069import org.apache.hadoop.hbase.master.assignment.RegionStates.RegionStateNode; -070import org.apache.hadoop.hbase.net.Address; -071import org.apache.hadoop.hbase.procedure2.Procedure; -072import org.apache.hadoop.hbase.protobuf.ProtobufMagic; -073import org.apache.hadoop.hbase.protobuf.ProtobufUtil; -074import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos; -075import org.apache.hadoop.hbase.protobuf.generated.RSGroupProtos; -076import org.apache.hadoop.hbase.regionserver.DisabledRegionSplitPolicy; -077import org.apache.hadoop.hbase.security.access.AccessControlLists; -078import org.apache.hadoop.hbase.util.Bytes; -079import org.apache.hadoop.hbase.zookeeper.ZKUtil; -080import org.apache.hadoop.hbase.zookeeper.ZKWatcher; -081import org.apache.hadoop.hbase.zookeeper.ZNodePaths; -082import org.apache.yetus.audience.InterfaceAudience; -083import org.apache.zookeeper.KeeperException; -084import org.slf4j.Logger; -085import org.slf4j.LoggerFactory; -086 -087import org.apache.hbase.thirdparty.com.google.common.collect.Lists; -088import org.apache.hbase.thirdparty.com.google.common.collect.Maps; -089import org.apache.hbase.thirdparty.com.google.common.collect.Sets; -090import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter; -091import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos; -092 -093/** -094 * This is an implementation of {@link RSGroupInfoManager} which makes -095 * use of an HBase table as the persistence store for the group information. -096 * It also makes use of zookeeper to store group information needed -097 * for bootstrapping during offline mode. -098 * -099 * h2Concurrency/h2 -100 * RSGroup state is kept locally in Maps. There is a rsgroup name to cached -101 * RSGroupInfo Map at {@link #rsGroupMap} and a Map of tables to the name of the -102 * rsgroup they belong too (in {@link #tableMap}). These Maps are persisted to the -103 * hbase:rsgroup table (and cached in zk) on each modification. -104 * -105 * pMutations on state are synchronized but reads can continue without having -106 * to wait on an instance monitor, mutations do wholesale replace of the Maps on -107 * update -- Copy-On-Write; the local Maps of state are read-only, just-in-case -108 * (see flushConfig). -109 * -110 * pReads must not block else there is a danger we'll deadlock. -111 * -112 * pClients of this class, the {@link RSGroupAdminEndpoint} for example, want to query and -113 * then act on the results of the query modifying cache in zookeeper without another thread -114 * making intermediate modifications. These clients synchronize on the 'this' instance so -115 * no other has access concurrently. Reads must be able to continue concurrently. -116 */ -117@InterfaceAudience.Private -118final class RSGroupInfoManagerImpl implements RSGroupInfoManager { -119 private static final Logger LOG =
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html -- diff --git a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html index bf1e334..f55c3a6 100644 --- a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html +++ b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html @@ -7,7 +7,7 @@ - + Apache HBase - Exemplar for hbase-client archetype Reactor Dependency Convergence @@ -912,7 +912,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-01-11 + Last Published: 2018-01-12 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html -- diff --git a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html index bdccdc9..692a5be 100644 --- a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html +++ b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html @@ -7,7 +7,7 @@ - + Apache HBase - Exemplar for hbase-client archetype Dependency Information @@ -147,7 +147,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-01-11 + Last Published: 2018-01-12 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html -- diff --git a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html index 482ab8a..04acfbe 100644 --- a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html +++ b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html @@ -7,7 +7,7 @@ - + Apache HBase - Exemplar for hbase-client archetype Project Dependency Management @@ -810,7 +810,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-01-11 + Last Published: 2018-01-12 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html -- diff --git a/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html b/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html index 9bf6e8e..9d5eee1 100644 --- a/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html +++ b/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html @@ -7,7 +7,7 @@ - + Apache HBase - Exemplar for hbase-client archetype About @@ -119,7 +119,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-01-11 + Last Published: 2018-01-12 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html -- diff --git a/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html b/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html index 06a6c6f..3b2cf47 100644 --- a/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html +++ b/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html @@ -7,7 +7,7 @@ - + Apache HBase - Exemplar for hbase-client archetype CI Management @@ -126,7 +126,7 @@
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f183e80f/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.DisabledWAL.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.DisabledWAL.html b/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.DisabledWAL.html index 0b621b6..69a8636 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.DisabledWAL.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.DisabledWAL.html @@ -82,7 +82,7 @@ 074 } 075 076 @Override -077 public WAL getWAL(final byte[] identifier, byte[] namespace) throws IOException { +077 public WAL getWAL(RegionInfo region) throws IOException { 078return disabled; 079 } 080 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f183e80f/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.html b/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.html index 0b621b6..69a8636 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/wal/DisabledWALProvider.html @@ -82,7 +82,7 @@ 074 } 075 076 @Override -077 public WAL getWAL(final byte[] identifier, byte[] namespace) throws IOException { +077 public WAL getWAL(RegionInfo region) throws IOException { 078return disabled; 079 } 080 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f183e80f/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.IdentityGroupingStrategy.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.IdentityGroupingStrategy.html b/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.IdentityGroupingStrategy.html index 4fec5eb..63c9ca7 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.IdentityGroupingStrategy.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/wal/RegionGroupingProvider.IdentityGroupingStrategy.html @@ -35,254 +35,262 @@ 027import java.util.List; 028import java.util.concurrent.ConcurrentHashMap; 029import java.util.concurrent.ConcurrentMap; -030 +030import java.util.concurrent.locks.Lock; 031import org.apache.hadoop.conf.Configuration; -032import org.apache.yetus.audience.InterfaceAudience; -033import org.slf4j.Logger; -034import org.slf4j.LoggerFactory; -035// imports for classes still in regionserver.wal -036import org.apache.hadoop.hbase.regionserver.wal.WALActionsListener; -037import org.apache.hadoop.hbase.util.Bytes; -038import org.apache.hadoop.hbase.util.IdLock; -039 -040/** -041 * A WAL Provider that returns a WAL per group of regions. -042 * -043 * This provider follows the decorator pattern and mainly holds the logic for WAL grouping. -044 * WAL creation/roll/close is delegated to {@link #DELEGATE_PROVIDER} -045 * -046 * Region grouping is handled via {@link RegionGroupingStrategy} and can be configured via the -047 * property "hbase.wal.regiongrouping.strategy". Current strategy choices are -048 * ul -049 * liemdefaultStrategy/em : Whatever strategy this version of HBase picks. currently -050 * "bounded"./li -051 * liemidentity/em : each region belongs to its own group./li -052 * liembounded/em : bounded number of groups and region evenly assigned to each group./li -053 * /ul -054 * Optionally, a FQCN to a custom implementation may be given. -055 */ -056@InterfaceAudience.Private -057public class RegionGroupingProvider implements WALProvider { -058 private static final Logger LOG = LoggerFactory.getLogger(RegionGroupingProvider.class); -059 -060 /** -061 * Map identifiers to a group number. -062 */ -063 public static interface RegionGroupingStrategy { -064String GROUP_NAME_DELIMITER = "."; -065 -066/** -067 * Given an identifier and a namespace, pick a group. -068 */ -069String group(final byte[] identifier, byte[] namespace); -070void init(Configuration config, String providerId); -071 } -072 -073 /** -074 * Maps between configuration names for strategies and implementation classes. -075 */ -076 static enum Strategies { -077 defaultStrategy(BoundedGroupingStrategy.class), -078 identity(IdentityGroupingStrategy.class), -079 bounded(BoundedGroupingStrategy.class), -080 namespace(NamespaceGroupingStrategy.class); -081 -082final Class? extends RegionGroupingStrategy clazz; -083Strategies(Class? extends RegionGroupingStrategy clazz) { -084 this.clazz = clazz; -085} -086 } -087 -088 /** -089 * instantiate a strategy from a config
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html index 67417b5..900808a 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html +++ b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AbortProcedureFuture.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -private static class HBaseAdmin.AbortProcedureFuture +private static class HBaseAdmin.AbortProcedureFuture extends HBaseAdmin.ProcedureFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean @@ -235,7 +235,7 @@ extends isAbortInProgress -privateboolean isAbortInProgress +privateboolean isAbortInProgress @@ -252,7 +252,7 @@ extends AbortProcedureFuture -publicAbortProcedureFuture(HBaseAdminadmin, +publicAbortProcedureFuture(HBaseAdminadmin, http://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">LongprocId, http://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">BooleanabortProcResponse) @@ -271,7 +271,7 @@ extends get -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Booleanget(longtimeout, +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Booleanget(longtimeout, http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true; title="class or interface in java.util.concurrent">TimeUnitunit) throws http://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true; title="class or interface in java.lang">InterruptedException, http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutionException.html?is-external=true; title="class or interface in java.util.concurrent">ExecutionException, http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html index 4e91404..9d834f8 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html +++ b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.AddColumnFamilyFuture.html @@ -132,7 +132,7 @@ var activeTableTab = "activeTableTab"; -private static class HBaseAdmin.AddColumnFamilyFuture +private static class HBaseAdmin.AddColumnFamilyFuture extends HBaseAdmin.ModifyTableFuture @@ -246,7 +246,7 @@ extends AddColumnFamilyFuture -publicAddColumnFamilyFuture(HBaseAdminadmin, +publicAddColumnFamilyFuture(HBaseAdminadmin, TableNametableName, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponseresponse) @@ -265,7 +265,7 @@ extends getOperationType -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetOperationType() +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetOperationType() Overrides: getOperationTypein classHBaseAdmin.ModifyTableFuture http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html index 485fa7a..40023f1 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html +++ b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html @@ -127,7 +127,7 @@ var activeTableTab = "activeTableTab"; -private static class HBaseAdmin.CreateTableFuture +private static class HBaseAdmin.CreateTableFuture extends HBaseAdmin.TableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void @@ -268,7 +268,7 @@ extends desc -private finalTableDescriptor desc +private
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c7c40c62/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html index 19aa0b9..aedcf26 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html +++ b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html @@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -class RawAsyncHBaseAdmin +class RawAsyncHBaseAdmin extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements AsyncAdmin The implementation of AsyncAdmin. @@ -704,12 +704,12 @@ implements -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureClusterStatus -getClusterStatus() +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureClusterMetrics +getClusterMetrics() -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureClusterStatus -getClusterStatus(http://docs.oracle.com/javase/8/docs/api/java/util/EnumSet.html?is-external=true; title="class or interface in java.util">EnumSetClusterMetrics.Optionoptions) +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureClusterMetrics +getClusterMetrics(http://docs.oracle.com/javase/8/docs/api/java/util/EnumSet.html?is-external=true; title="class or interface in java.util">EnumSetClusterMetrics.Optionoptions) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureCompactionState @@ -788,27 +788,27 @@ implements -private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionLoad -getRegionLoad(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequestrequest, - ServerNameserverName) +(package private) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureHRegionLocation +getRegionLocation(byte[]regionNameOrEncodedRegionName) +Get the region location for the passed region name. + -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionLoad -getRegionLoads(ServerNameserverName) -Get a list of RegionLoad of all regions hosted on a region seerver. - +private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionMetrics +getRegionMetrics(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequestrequest, +ServerNameserverName) -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionLoad -getRegionLoads(ServerNameserverName, - TableNametableName) -Get a list of RegionLoad of all regions hosted on a region seerver for a table. +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionMetrics +getRegionMetrics(ServerNameserverName) +Get a list of RegionMetrics of all regions hosted on a region seerver. -(package private) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bb398572/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html index 5b3b750..a1f3f7e 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html @@ -97,3307 +97,3304 @@ 089import org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer; 090import org.apache.hbase.thirdparty.io.netty.util.Timeout; 091import org.apache.hbase.thirdparty.io.netty.util.TimerTask; -092import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -093import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter; -094import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService; -095import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest; -096import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesResponse; -097import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest; -098import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse; -099import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest; -100import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionResponse; -101import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionRequest; -102import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionResponse; -103import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest; -104import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse; -105import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequest; -106import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse; -107import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest; -108import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse; -109import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest; -110import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerResponse; -111import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest; -112import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationResponse; -113import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription; -114import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType; -115import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema; -116import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest; -117import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse; -118import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest; -119import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse; -120import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest; -121import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse; -122import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest; -123import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse; -124import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest; -125import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse; -126import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest; -127import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse; -128import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest; -129import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse; -130import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest; -131import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse; -132import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest; -133import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse;
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html b/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html index 1fd770d..5c0a3d9 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html +++ b/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.ModifyableTableDescriptor.html @@ -1961,6 +1961,6 @@ publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.h -Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.html b/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.html index 1ce7c03..782e4e3 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.html +++ b/devapidocs/org/apache/hadoop/hbase/client/TableDescriptorBuilder.html @@ -1553,6 +1553,6 @@ publicCopyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html b/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html index 037cb4a..dd4afb1 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html +++ b/devapidocs/org/apache/hadoop/hbase/client/TableSnapshotScanner.html @@ -620,6 +620,6 @@ extends Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html b/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html index f65cdb2..66f3198 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html +++ b/devapidocs/org/apache/hadoop/hbase/client/TableState.State.html @@ -409,6 +409,6 @@ not permitted.) -Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/TableState.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/TableState.html b/devapidocs/org/apache/hadoop/hbase/client/TableState.html index 310a067..0ea4ab9 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/TableState.html +++ b/devapidocs/org/apache/hadoop/hbase/client/TableState.html @@ -556,6 +556,6 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? -Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.html b/devapidocs/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.html index 76cec51..6e021bc 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.html +++ b/devapidocs/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.html @@ -356,6 +356,6 @@ extends -Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/VersionInfoUtil.html
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/83bf6175/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionMapper.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionMapper.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionMapper.html index 91eec45..d1cd185 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionMapper.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/CompactionTool.CompactionMapper.html @@ -88,400 +88,396 @@ 080 private final static String CONF_COMPACT_ONCE = "hbase.compactiontool.compact.once"; 081 private final static String CONF_COMPACT_MAJOR = "hbase.compactiontool.compact.major"; 082 private final static String CONF_DELETE_COMPACTED = "hbase.compactiontool.delete"; -083 private final static String CONF_COMPLETE_COMPACTION = "hbase.hstore.compaction.complete"; -084 -085 /** -086 * Class responsible to execute the Compaction on the specified path. -087 * The path can be a table, region or family directory. -088 */ -089 private static class CompactionWorker { -090private final boolean keepCompactedFiles; -091private final boolean deleteCompacted; -092private final Configuration conf; -093private final FileSystem fs; -094private final Path tmpDir; -095 -096public CompactionWorker(final FileSystem fs, final Configuration conf) { -097 this.conf = conf; -098 this.keepCompactedFiles = !conf.getBoolean(CONF_COMPLETE_COMPACTION, true); -099 this.deleteCompacted = conf.getBoolean(CONF_DELETE_COMPACTED, false); -100 this.tmpDir = new Path(conf.get(CONF_TMP_DIR)); -101 this.fs = fs; -102} -103 -104/** -105 * Execute the compaction on the specified path. -106 * -107 * @param path Directory path on which to run compaction. -108 * @param compactOnce Execute just a single step of compaction. -109 * @param major Request major compaction. -110 */ -111public void compact(final Path path, final boolean compactOnce, final boolean major) throws IOException { -112 if (isFamilyDir(fs, path)) { -113Path regionDir = path.getParent(); -114Path tableDir = regionDir.getParent(); -115TableDescriptor htd = FSTableDescriptors.getTableDescriptorFromFs(fs, tableDir); -116RegionInfo hri = HRegionFileSystem.loadRegionInfoFileContent(fs, regionDir); -117compactStoreFiles(tableDir, htd, hri, -118path.getName(), compactOnce, major); -119 } else if (isRegionDir(fs, path)) { -120Path tableDir = path.getParent(); -121TableDescriptor htd = FSTableDescriptors.getTableDescriptorFromFs(fs, tableDir); -122compactRegion(tableDir, htd, path, compactOnce, major); -123 } else if (isTableDir(fs, path)) { -124compactTable(path, compactOnce, major); -125 } else { -126throw new IOException( -127 "Specified path is not a table, region or family directory. path=" + path); -128 } -129} -130 -131private void compactTable(final Path tableDir, final boolean compactOnce, final boolean major) -132throws IOException { -133 TableDescriptor htd = FSTableDescriptors.getTableDescriptorFromFs(fs, tableDir); -134 for (Path regionDir: FSUtils.getRegionDirs(fs, tableDir)) { -135compactRegion(tableDir, htd, regionDir, compactOnce, major); -136 } -137} -138 -139private void compactRegion(final Path tableDir, final TableDescriptor htd, -140final Path regionDir, final boolean compactOnce, final boolean major) -141throws IOException { -142 RegionInfo hri = HRegionFileSystem.loadRegionInfoFileContent(fs, regionDir); -143 for (Path familyDir: FSUtils.getFamilyDirs(fs, regionDir)) { -144compactStoreFiles(tableDir, htd, hri, familyDir.getName(), compactOnce, major); -145 } -146} -147 -148/** -149 * Execute the actual compaction job. -150 * If the compact once flag is not specified, execute the compaction until -151 * no more compactions are needed. Uses the Configuration settings provided. -152 */ -153private void compactStoreFiles(final Path tableDir, final TableDescriptor htd, -154final RegionInfo hri, final String familyName, final boolean compactOnce, -155final boolean major) throws IOException { -156 HStore store = getStore(conf, fs, tableDir, htd, hri, familyName, tmpDir); -157 LOG.info("Compact table=" + htd.getTableName() + -158" region=" + hri.getRegionNameAsString() + -159" family=" + familyName); -160 if (major) { -161store.triggerMajorCompaction(); -162 } -163 do { -164OptionalCompactionContext compaction = -165
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html b/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html index edd369a..6fa4fbd 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html +++ b/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html @@ -175,32 +175,32 @@ implements org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse -abortProcedure(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, +abortProcedure(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequestrequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse -addColumn(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, +addColumn(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequestrequest) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse -addReplicationPeer(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, +addReplicationPeer(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequestrequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse -assignRegion(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, +assignRegion(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequestrequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse -balance(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, +balance(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequestrequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse -clearDeadServers(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, +clearDeadServers(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequestrequest) @@ -209,347 +209,347 @@ implements org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse -createNamespace(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, +createNamespace(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequestrequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse -createTable(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, +createTable(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequestrequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse -decommissionRegionServers(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, +decommissionRegionServers(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequestrequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse -deleteColumn(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, +deleteColumn(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequestrequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse -deleteNamespace(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, +deleteNamespace(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequestrequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d449e87f/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.TagRewriteByteBufferExtendedCell.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.TagRewriteByteBufferExtendedCell.html b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.TagRewriteByteBufferExtendedCell.html new file mode 100644 index 000..662952e --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.TagRewriteByteBufferExtendedCell.html @@ -0,0 +1,1117 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +PrivateCellUtil.TagRewriteByteBufferExtendedCell (Apache HBase 3.0.0-SNAPSHOT API) + + + + + +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Field| +Constr| +Method + + +Detail: +Field| +Constr| +Method + + + + + + + + +org.apache.hadoop.hbase +Class PrivateCellUtil.TagRewriteByteBufferExtendedCell + + + +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.ByteBufferExtendedCell + + +org.apache.hadoop.hbase.PrivateCellUtil.TagRewriteByteBufferExtendedCell + + + + + + + + + +All Implemented Interfaces: +http://docs.oracle.com/javase/8/docs/api/java/lang/Cloneable.html?is-external=true; title="class or interface in java.lang">Cloneable, Cell, ExtendedCell, HeapSize, RawCell + + +Direct Known Subclasses: +PrivateCellUtil.ValueAndTagRewriteByteBufferExtendedCell + + +Enclosing class: +PrivateCellUtil + + + +static class PrivateCellUtil.TagRewriteByteBufferExtendedCell +extends ByteBufferExtendedCell + + + + + + + + + + + +Nested Class Summary + + + + +Nested classes/interfaces inherited from interfaceorg.apache.hadoop.hbase.Cell +Cell.Type + + + + + + + + +Field Summary + +Fields + +Modifier and Type +Field and Description + + +protected ByteBufferExtendedCell +cell + + +private static int +HEAP_SIZE_OVERHEAD + + +protected byte[] +tags + + + + + + +Fields inherited from interfaceorg.apache.hadoop.hbase.ExtendedCell +CELL_NOT_BASED_ON_CHUNK + + + + + +Fields inherited from interfaceorg.apache.hadoop.hbase.RawCell +MAX_TAGS_LENGTH + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +TagRewriteByteBufferExtendedCell(ByteBufferExtendedCellcell, +byte[]tags) + + + + + + + + + +Method Summary + +All MethodsInstance MethodsConcrete Methods + +Modifier and Type +Method and Description + + +ExtendedCell +deepClone() +Does a deep copy of the contents to a new memory area and returns it as a new cell. + + + +byte[] +getFamilyArray() +Contiguous bytes composed of legal HDFS filename characters which may start at any index in the + containing array. + + + +http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true; title="class or interface in java.nio">ByteBuffer +getFamilyByteBuffer() + + +byte +getFamilyLength() + + +int +getFamilyOffset() + + +int +getFamilyPosition() + + +byte[] +getQualifierArray() +Contiguous raw bytes that may start at any index in the containing array. + + + +http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true; title="class or interface in java.nio">ByteBuffer +getQualifierByteBuffer() + + +int +getQualifierLength() + + +int +getQualifierOffset() + + +int +getQualifierPosition() + + +byte[] +getRowArray() +Contiguous raw bytes that may start at any index in the containing array. + + + +http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true; title="class or interface in java.nio">ByteBuffer +getRowByteBuffer() + + +short +getRowLength() + + +int +getRowOffset() + + +int +getRowPosition() + +
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d2b28a1a/devapidocs/org/apache/hadoop/hbase/client/example/HttpProxyExample.RequestHandler.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/example/HttpProxyExample.RequestHandler.html b/devapidocs/org/apache/hadoop/hbase/client/example/HttpProxyExample.RequestHandler.html index 3d37d41..5a14666 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/example/HttpProxyExample.RequestHandler.html +++ b/devapidocs/org/apache/hadoop/hbase/client/example/HttpProxyExample.RequestHandler.html @@ -132,7 +132,7 @@ var activeTableTab = "activeTableTab"; -private static final class HttpProxyExample.RequestHandler +private static final class HttpProxyExample.RequestHandler extends org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHandlerorg.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.FullHttpRequest @@ -302,7 +302,7 @@ extends org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand conn -private finalAsyncConnection conn +private finalAsyncConnection conn @@ -311,7 +311,7 @@ extends org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand channelGroup -private finalorg.apache.hadoop.hbase.shaded.io.netty.channel.group.ChannelGroup channelGroup +private finalorg.apache.hadoop.hbase.shaded.io.netty.channel.group.ChannelGroup channelGroup @@ -328,7 +328,7 @@ extends org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand RequestHandler -publicRequestHandler(AsyncConnectionconn, +publicRequestHandler(AsyncConnectionconn, org.apache.hadoop.hbase.shaded.io.netty.channel.group.ChannelGroupchannelGroup) @@ -346,7 +346,7 @@ extends org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand channelActive -publicvoidchannelActive(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx) +publicvoidchannelActive(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx) Specified by: channelActivein interfaceorg.apache.hadoop.hbase.shaded.io.netty.channel.ChannelInboundHandler @@ -361,7 +361,7 @@ extends org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand channelInactive -publicvoidchannelInactive(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx) +publicvoidchannelInactive(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx) Specified by: channelInactivein interfaceorg.apache.hadoop.hbase.shaded.io.netty.channel.ChannelInboundHandler @@ -376,7 +376,7 @@ extends org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand write -privatevoidwrite(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx, +privatevoidwrite(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx, org.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.HttpResponseStatusstatus, http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">Optionalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringcontent) @@ -387,7 +387,7 @@ extends org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand parse -privateHttpProxyExample.Paramsparse(org.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.FullHttpRequestreq) +privateHttpProxyExample.Paramsparse(org.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.FullHttpRequestreq) @@ -396,7 +396,7 @@ extends org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand get -privatevoidget(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx, +privatevoidget(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx, org.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.FullHttpRequestreq) @@ -406,7 +406,7 @@ extends org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand put -privatevoidput(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx, +privatevoidput(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx, org.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.FullHttpRequestreq) @@ -416,7 +416,7 @@ extends org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHand channelRead0 -protectedvoidchannelRead0(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx, +protectedvoidchannelRead0(org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContextctx, org.apache.hadoop.hbase.shaded.io.netty.handler.codec.http.FullHttpRequestreq) Specified by: @@ -430,7 +430,7 @@ extends
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b618ac40/devapidocs/org/apache/hadoop/hbase/master/HMaster.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html index 0c57dc7..061085c 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html +++ b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html @@ -128,7 +128,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.LimitedPrivate(value="Tools") -public class HMaster +public class HMaster extends HRegionServer implements MasterServices HMaster is the "master server" for HBase. An HBase cluster has one active @@ -1434,7 +1434,7 @@ implements LOG -private staticorg.slf4j.Logger LOG +private staticorg.slf4j.Logger LOG @@ -1443,7 +1443,7 @@ implements MASTER -public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String MASTER +public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String MASTER See Also: Constant Field Values @@ -1456,7 +1456,7 @@ implements activeMasterManager -private finalActiveMasterManager activeMasterManager +private finalActiveMasterManager activeMasterManager @@ -1465,7 +1465,7 @@ implements regionServerTracker -RegionServerTracker regionServerTracker +RegionServerTracker regionServerTracker @@ -1474,7 +1474,7 @@ implements drainingServerTracker -privateDrainingServerTracker drainingServerTracker +privateDrainingServerTracker drainingServerTracker @@ -1483,7 +1483,7 @@ implements loadBalancerTracker -LoadBalancerTracker loadBalancerTracker +LoadBalancerTracker loadBalancerTracker @@ -1492,7 +1492,7 @@ implements splitOrMergeTracker -privateSplitOrMergeTracker splitOrMergeTracker +privateSplitOrMergeTracker splitOrMergeTracker @@ -1501,7 +1501,7 @@ implements regionNormalizerTracker -privateRegionNormalizerTracker regionNormalizerTracker +privateRegionNormalizerTracker regionNormalizerTracker @@ -1510,7 +1510,7 @@ implements maintenanceModeTracker -privateMasterMaintenanceModeTracker maintenanceModeTracker +privateMasterMaintenanceModeTracker maintenanceModeTracker @@ -1519,7 +1519,7 @@ implements clusterSchemaService -privateClusterSchemaService clusterSchemaService +privateClusterSchemaService clusterSchemaService @@ -1528,7 +1528,7 @@ implements HBASE_MASTER_WAIT_ON_SERVICE_IN_SECONDS -public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String HBASE_MASTER_WAIT_ON_SERVICE_IN_SECONDS +public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String HBASE_MASTER_WAIT_ON_SERVICE_IN_SECONDS See Also: Constant Field Values @@ -1541,7 +1541,7 @@ implements DEFAULT_HBASE_MASTER_WAIT_ON_SERVICE_IN_SECONDS -public static finalint DEFAULT_HBASE_MASTER_WAIT_ON_SERVICE_IN_SECONDS +public static finalint DEFAULT_HBASE_MASTER_WAIT_ON_SERVICE_IN_SECONDS See Also: Constant Field Values @@ -1554,7 +1554,7 @@ implements metricsMaster -finalMetricsMaster metricsMaster +finalMetricsMaster metricsMaster @@ -1563,7 +1563,7 @@ implements fileSystemManager -privateMasterFileSystem fileSystemManager +privateMasterFileSystem fileSystemManager @@ -1572,7 +1572,7 @@ implements walManager -privateMasterWalManager walManager +privateMasterWalManager walManager @@ -1581,7 +1581,7 @@ implements serverManager -private volatileServerManager serverManager +private volatileServerManager serverManager @@ -1590,7 +1590,7 @@ implements assignmentManager -privateAssignmentManager assignmentManager +privateAssignmentManager assignmentManager @@ -1599,7 +1599,7 @@ implements replicationManager -privateReplicationManager replicationManager +privateReplicationManager replicationManager @@ -1608,7 +1608,7 @@ implements rsFatals -MemoryBoundedLogMessageBuffer rsFatals +MemoryBoundedLogMessageBuffer rsFatals @@ -1617,7 +1617,7 @@ implements activeMaster -private volatileboolean activeMaster +private volatileboolean activeMaster @@ -1626,7 +1626,7 @@ implements initialized -private finalProcedureEvent initialized +private finalProcedureEvent initialized @@ -1635,7 +1635,7 @@ implements serviceStarted -volatileboolean serviceStarted +volatileboolean serviceStarted @@ -1644,7 +1644,7 @@ implements serverCrashProcessingEnabled -private finalProcedureEvent serverCrashProcessingEnabled +private finalProcedureEvent serverCrashProcessingEnabled @@ -1653,7 +1653,7 @@ implements maxBlancingTime -private finalint maxBlancingTime +private
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/org/apache/hadoop/hbase/security/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/security/package-tree.html b/devapidocs/org/apache/hadoop/hbase/security/package-tree.html index c4d3f6e..bec1bbc 100644 --- a/devapidocs/org/apache/hadoop/hbase/security/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/security/package-tree.html @@ -191,8 +191,8 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection org.apache.hadoop.hbase.security.AuthMethod +org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection org.apache.hadoop.hbase.security.SaslStatus http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html index cd08056..570c4fa 100644 --- a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html @@ -198,9 +198,9 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) +org.apache.hadoop.hbase.thrift.ThriftMetrics.ThriftServerType org.apache.hadoop.hbase.thrift.ThriftServerRunner.ImplType org.apache.hadoop.hbase.thrift.MetricsThriftServerSourceFactoryImpl.FactoryStorage -org.apache.hadoop.hbase.thrift.ThriftMetrics.ThriftServerType http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html b/devapidocs/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html index 43c904b..58a529e 100644 --- a/devapidocs/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html +++ b/devapidocs/org/apache/hadoop/hbase/util/ServerRegionReplicaUtil.html @@ -320,7 +320,7 @@ extends Admin java class. Replication to secondary region replicas works over standard inter-cluster replication.· See Also: http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/org/apache/hadoop/hbase/util/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html index bc6e3ef..9aba27a 100644 --- a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html @@ -536,14 +536,14 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.util.ChecksumType -org.apache.hadoop.hbase.util.IdReadWriteLock.ReferenceType -org.apache.hadoop.hbase.util.PrettyPrinter.Unit org.apache.hadoop.hbase.util.Order -org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE -org.apache.hadoop.hbase.util.PoolMap.PoolType +org.apache.hadoop.hbase.util.ChecksumType org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.UnsafeComparer (implements org.apache.hadoop.hbase.util.Bytes.ComparerT) +org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.PureJavaComparer (implements org.apache.hadoop.hbase.util.Bytes.ComparerT) +org.apache.hadoop.hbase.util.PrettyPrinter.Unit +org.apache.hadoop.hbase.util.IdReadWriteLock.ReferenceType +org.apache.hadoop.hbase.util.PoolMap.PoolType
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4b2cc17/apidocs/src-html/org/apache/hadoop/hbase/client/Admin.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/client/Admin.html b/apidocs/src-html/org/apache/hadoop/hbase/client/Admin.html index d7f2c1a..9b468f8 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/client/Admin.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/client/Admin.html @@ -36,2607 +36,2606 @@ 028import java.util.TreeMap; 029import java.util.concurrent.Future; 030import java.util.regex.Pattern; -031 -032import org.apache.hadoop.conf.Configuration; -033import org.apache.hadoop.hbase.Abortable; -034import org.apache.hadoop.hbase.CacheEvictionStats; +031import org.apache.hadoop.conf.Configuration; +032import org.apache.hadoop.hbase.Abortable; +033import org.apache.hadoop.hbase.CacheEvictionStats; +034import org.apache.hadoop.hbase.ClusterMetrics.Option; 035import org.apache.hadoop.hbase.ClusterStatus; -036import org.apache.hadoop.hbase.ClusterStatus.Option; -037import org.apache.hadoop.hbase.HRegionInfo; -038import org.apache.hadoop.hbase.HTableDescriptor; -039import org.apache.hadoop.hbase.NamespaceDescriptor; -040import org.apache.hadoop.hbase.NamespaceNotFoundException; -041import org.apache.hadoop.hbase.RegionLoad; -042import org.apache.hadoop.hbase.ServerName; -043import org.apache.hadoop.hbase.TableExistsException; -044import org.apache.hadoop.hbase.TableName; -045import org.apache.hadoop.hbase.TableNotFoundException; -046import org.apache.yetus.audience.InterfaceAudience; -047import org.apache.hadoop.hbase.client.replication.TableCFs; -048import org.apache.hadoop.hbase.client.security.SecurityCapability; -049import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel; -050import org.apache.hadoop.hbase.quotas.QuotaFilter; -051import org.apache.hadoop.hbase.quotas.QuotaRetriever; -052import org.apache.hadoop.hbase.quotas.QuotaSettings; -053import org.apache.hadoop.hbase.regionserver.wal.FailedLogCloseException; -054import org.apache.hadoop.hbase.replication.ReplicationException; -055import org.apache.hadoop.hbase.replication.ReplicationPeerConfig; -056import org.apache.hadoop.hbase.replication.ReplicationPeerDescription; -057import org.apache.hadoop.hbase.snapshot.HBaseSnapshotException; -058import org.apache.hadoop.hbase.snapshot.RestoreSnapshotException; -059import org.apache.hadoop.hbase.snapshot.SnapshotCreationException; -060import org.apache.hadoop.hbase.snapshot.UnknownSnapshotException; -061import org.apache.hadoop.hbase.util.Bytes; -062import org.apache.hadoop.hbase.util.Pair; -063 -064/** -065 * The administrative API for HBase. Obtain an instance from {@link Connection#getAdmin()} and -066 * call {@link #close()} when done. -067 * pAdmin can be used to create, drop, list, enable and disable and otherwise modify tables, -068 * as well as perform other administrative operations. -069 * -070 * @see ConnectionFactory -071 * @see Connection -072 * @see Table -073 * @since 0.99.0 -074 */ -075@InterfaceAudience.Public -076public interface Admin extends Abortable, Closeable { -077 int getOperationTimeout(); -078 -079 @Override -080 void abort(String why, Throwable e); -081 -082 @Override -083 boolean isAborted(); -084 -085 /** -086 * @return Connection used by this object. -087 */ -088 Connection getConnection(); -089 -090 /** -091 * @param tableName Table to check. -092 * @return codetrue/code if table exists already. -093 * @throws IOException -094 */ -095 boolean tableExists(TableName tableName) throws IOException; -096 -097 /** -098 * List all the userspace tables. -099 * -100 * @return an array of read-only HTableDescriptors -101 * @throws IOException if a remote or network exception occurs -102 * @deprecated since 2.0 version and will be removed in 3.0 version. -103 * Use {@link #listTableDescriptors()}. -104 * @see #listTableDescriptors() -105 */ -106 @Deprecated -107 HTableDescriptor[] listTables() throws IOException; -108 -109 /** -110 * List all the userspace tables. -111 * -112 * @return a list of TableDescriptors -113 * @throws IOException if a remote or network exception occurs -114 */ -115 ListTableDescriptor listTableDescriptors() throws IOException; -116 -117 /** -118 * List all the userspace tables that match the given pattern. -119 * -120 * @param pattern The compiled regular expression to match against -121 * @return an array of read-only HTableDescriptors -122 * @throws IOException if a remote or network exception occurs -123 * @see #listTables() -124 * @deprecated since 2.0 version and will be removed in 3.0 version. -125 * Use {@link #listTableDescriptors(java.util.regex.Pattern)}. -126 * @see #listTableDescriptors(Pattern) -127 */ -128 @Deprecated -129 HTableDescriptor[] listTables(Pattern pattern) throws
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/505bbb2e/apidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotInfo.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotInfo.html b/apidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotInfo.html index 6c97088..f999cbd 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotInfo.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/snapshot/SnapshotInfo.html @@ -42,679 +42,678 @@ 034 035import org.apache.commons.cli.CommandLine; 036import org.apache.commons.cli.Option; -037import org.apache.commons.logging.Log; -038import org.apache.commons.logging.LogFactory; -039import org.apache.hadoop.conf.Configuration; -040import org.apache.hadoop.fs.FileStatus; -041import org.apache.hadoop.fs.FileSystem; -042import org.apache.hadoop.fs.Path; -043import org.apache.hadoop.hbase.TableName; -044import org.apache.hadoop.hbase.client.RegionInfo; -045import org.apache.hadoop.hbase.client.SnapshotDescription; -046import org.apache.hadoop.hbase.io.HFileLink; -047import org.apache.hadoop.hbase.io.WALLink; -048import org.apache.hadoop.hbase.util.AbstractHBaseTool; -049import org.apache.hadoop.hbase.util.FSUtils; -050import org.apache.hadoop.util.StringUtils; -051import org.apache.yetus.audience.InterfaceAudience; -052 -053import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -054import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos; -055import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotRegionManifest; -056 -057/** -058 * Tool for dumping snapshot information. -059 * ol -060 * li Table Descriptor -061 * li Snapshot creation time, type, format version, ... -062 * li List of hfiles and wals -063 * li Stats about hfiles and logs sizes, percentage of shared with the source table, ... -064 * /ol -065 */ -066@InterfaceAudience.Public -067public final class SnapshotInfo extends AbstractHBaseTool { -068 private static final Log LOG = LogFactory.getLog(SnapshotInfo.class); -069 -070 static final class Options { -071static final Option SNAPSHOT = new Option(null, "snapshot", true, "Snapshot to examine."); -072static final Option REMOTE_DIR = new Option(null, "remote-dir", true, -073"Root directory that contains the snapshots."); -074static final Option LIST_SNAPSHOTS = new Option(null, "list-snapshots", false, -075"List all the available snapshots and exit."); -076static final Option FILES = new Option(null, "files", false, "Files and logs list."); -077static final Option STATS = new Option(null, "stats", false, "Files and logs stats."); -078static final Option SCHEMA = new Option(null, "schema", false, -079"Describe the snapshotted table."); -080static final Option SIZE_IN_BYTES = new Option(null, "size-in-bytes", false, -081"Print the size of the files in bytes."); -082 } -083 -084 /** -085 * Statistics about the snapshot -086 * ol -087 * li How many store files and logs are in the archive -088 * li How many store files and logs are shared with the table -089 * li Total store files and logs size and shared amount -090 * /ol -091 */ -092 public static class SnapshotStats { -093/** Information about the file referenced by the snapshot */ -094static class FileInfo { -095 private final boolean corrupted; -096 private final boolean inArchive; -097 private final long size; -098 -099 FileInfo(final boolean inArchive, final long size, final boolean corrupted) { -100this.corrupted = corrupted; -101this.inArchive = inArchive; -102this.size = size; -103 } -104 -105 /** @return true if the file is in the archive */ -106 public boolean inArchive() { -107return this.inArchive; -108 } -109 -110 /** @return true if the file is corrupted */ -111 public boolean isCorrupted() { -112return this.corrupted; -113 } -114 -115 /** @return true if the file is missing */ -116 public boolean isMissing() { -117return this.size 0; -118 } -119 -120 /** @return the file size */ -121 public long getSize() { -122return this.size; -123 } -124 -125 String getStateToString() { -126if (isCorrupted()) return "CORRUPTED"; -127if (isMissing()) return "NOT FOUND"; -128if (inArchive()) return "archive"; -129return null; -130 } -131} -132 -133private AtomicInteger hfilesArchiveCount = new AtomicInteger(); -134private AtomicInteger hfilesCorrupted = new AtomicInteger(); -135private AtomicInteger hfilesMissing = new AtomicInteger(); -136private AtomicInteger hfilesCount = new AtomicInteger(); -137private AtomicInteger hfilesMobCount = new AtomicInteger(); -138private AtomicInteger logsMissing = new
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d0f1a9f6/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html index e354b98..d875570 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html @@ -129,117 +129,116 @@ 121 */ 122 @Override 123 public void execute() throws IOException { -124try (Admin admin = conn.getAdmin();) { -125 -126 // Begin BACKUP -127 beginBackup(backupManager, backupInfo); -128 String savedStartCode = null; -129 boolean firstBackup = false; -130 // do snapshot for full table backup -131 -132 savedStartCode = backupManager.readBackupStartCode(); -133 firstBackup = savedStartCode == null || Long.parseLong(savedStartCode) == 0L; -134 if (firstBackup) { -135// This is our first backup. Let's put some marker to system table so that we can hold the logs -136// while we do the backup. -137 backupManager.writeBackupStartCode(0L); -138 } -139 // We roll log here before we do the snapshot. It is possible there is duplicate data -140 // in the log that is already in the snapshot. But if we do it after the snapshot, we -141 // could have data loss. -142 // A better approach is to do the roll log on each RS in the same global procedure as -143 // the snapshot. -144 LOG.info("Execute roll log procedure for full backup ..."); -145 -146 MapString, String props = new HashMapString, String(); -147 props.put("backupRoot", backupInfo.getBackupRootDir()); -148 admin.execProcedure(LogRollMasterProcedureManager.ROLLLOG_PROCEDURE_SIGNATURE, -149 LogRollMasterProcedureManager.ROLLLOG_PROCEDURE_NAME, props); -150 -151 newTimestamps = backupManager.readRegionServerLastLogRollResult(); -152 if (firstBackup) { -153// Updates registered log files -154// We record ALL old WAL files as registered, because -155// this is a first full backup in the system and these -156// files are not needed for next incremental backup -157ListString logFiles = BackupUtils.getWALFilesOlderThan(conf, newTimestamps); -158 backupManager.recordWALFiles(logFiles); -159 } -160 -161 // SNAPSHOT_TABLES: -162 backupInfo.setPhase(BackupPhase.SNAPSHOT); -163 for (TableName tableName : tableList) { -164String snapshotName = -165"snapshot_" + Long.toString(EnvironmentEdgeManager.currentTime()) + "_" -166+ tableName.getNamespaceAsString() + "_" + tableName.getQualifierAsString(); -167 -168snapshotTable(admin, tableName, snapshotName); -169 backupInfo.setSnapshotName(tableName, snapshotName); -170 } -171 -172 // SNAPSHOT_COPY: -173 // do snapshot copy -174 LOG.debug("snapshot copy for " + backupId); -175 snapshotCopy(backupInfo); -176 // Updates incremental backup table set -177 backupManager.addIncrementalBackupTableSet(backupInfo.getTables()); -178 -179 // BACKUP_COMPLETE: -180 // set overall backup status: complete. Here we make sure to complete the backup. -181 // After this checkpoint, even if entering cancel process, will let the backup finished -182 backupInfo.setState(BackupState.COMPLETE); -183 // The table list in backupInfo is good for both full backup and incremental backup. -184 // For incremental backup, it contains the incremental backup table set. -185 backupManager.writeRegionServerLogTimestamp(backupInfo.getTables(), newTimestamps); -186 -187 HashMapTableName, HashMapString, Long newTableSetTimestampMap = -188 backupManager.readLogTimestampMap(); -189 -190 Long newStartCode = -191 BackupUtils.getMinValue(BackupUtils -192 .getRSLogTimestampMins(newTableSetTimestampMap)); -193 backupManager.writeBackupStartCode(newStartCode); -194 -195 // backup complete -196 completeBackup(conn, backupInfo, backupManager, BackupType.FULL, conf); -197} catch (Exception e) { -198 failBackup(conn, backupInfo, backupManager, e, "Unexpected BackupException : ", -199BackupType.FULL, conf); -200 throw new IOException(e); -201} -202 -203 } +124try (Admin admin = conn.getAdmin()) { +125 // Begin BACKUP +126 beginBackup(backupManager, backupInfo); +127 String savedStartCode = null; +128 boolean firstBackup = false; +129 // do snapshot for full table backup +130 +131 savedStartCode = backupManager.readBackupStartCode(); +132 firstBackup = savedStartCode ==
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4c0cfa5/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html b/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html index 4dedbc2..ffeadbf 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html @@ -902,350 +902,351 @@ 894return allowPartialResults; 895 } 896 -897 public Scan setLoadColumnFamiliesOnDemand(boolean value) { -898return (Scan) super.setLoadColumnFamiliesOnDemand(value); -899 } -900 -901 /** -902 * Compile the table and column family (i.e. schema) information -903 * into a String. Useful for parsing and aggregation by debugging, -904 * logging, and administration tools. -905 * @return Map -906 */ -907 @Override -908 public MapString, Object getFingerprint() { -909MapString, Object map = new HashMap(); -910ListString families = new ArrayList(); -911if(this.familyMap.isEmpty()) { -912 map.put("families", "ALL"); -913 return map; -914} else { -915 map.put("families", families); -916} -917for (Map.Entrybyte [], NavigableSetbyte[] entry : -918this.familyMap.entrySet()) { -919 families.add(Bytes.toStringBinary(entry.getKey())); -920} -921return map; -922 } -923 -924 /** -925 * Compile the details beyond the scope of getFingerprint (row, columns, -926 * timestamps, etc.) into a Map along with the fingerprinted information. -927 * Useful for debugging, logging, and administration tools. -928 * @param maxCols a limit on the number of columns output prior to truncation -929 * @return Map -930 */ -931 @Override -932 public MapString, Object toMap(int maxCols) { -933// start with the fingerpring map and build on top of it -934MapString, Object map = getFingerprint(); -935// map from families to column list replaces fingerprint's list of families -936MapString, ListString familyColumns = new HashMap(); -937map.put("families", familyColumns); -938// add scalar information first -939map.put("startRow", Bytes.toStringBinary(this.startRow)); -940map.put("stopRow", Bytes.toStringBinary(this.stopRow)); -941map.put("maxVersions", this.maxVersions); -942map.put("batch", this.batch); -943map.put("caching", this.caching); -944map.put("maxResultSize", this.maxResultSize); -945map.put("cacheBlocks", this.cacheBlocks); -946map.put("loadColumnFamiliesOnDemand", this.loadColumnFamiliesOnDemand); -947ListLong timeRange = new ArrayList(2); -948timeRange.add(this.tr.getMin()); -949timeRange.add(this.tr.getMax()); -950map.put("timeRange", timeRange); -951int colCount = 0; -952// iterate through affected families and list out up to maxCols columns -953for (Map.Entrybyte [], NavigableSetbyte[] entry : -954 this.familyMap.entrySet()) { -955 ListString columns = new ArrayList(); -956 familyColumns.put(Bytes.toStringBinary(entry.getKey()), columns); -957 if(entry.getValue() == null) { -958colCount++; -959--maxCols; -960columns.add("ALL"); -961 } else { -962colCount += entry.getValue().size(); -963if (maxCols = 0) { -964 continue; -965} -966for (byte [] column : entry.getValue()) { -967 if (--maxCols = 0) { -968continue; -969 } -970 columns.add(Bytes.toStringBinary(column)); -971} -972 } -973} -974map.put("totalColumns", colCount); -975if (this.filter != null) { -976 map.put("filter", this.filter.toString()); -977} -978// add the id if set -979if (getId() != null) { -980 map.put("id", getId()); -981} -982return map; -983 } -984 -985 /** -986 * Enable/disable "raw" mode for this scan. -987 * If "raw" is enabled the scan will return all -988 * delete marker and deleted rows that have not -989 * been collected, yet. -990 * This is mostly useful for Scan on column families -991 * that have KEEP_DELETED_ROWS enabled. -992 * It is an error to specify any column when "raw" is set. -993 * @param raw True/False to enable/disable "raw" mode. -994 */ -995 public Scan setRaw(boolean raw) { -996setAttribute(RAW_ATTR, Bytes.toBytes(raw)); -997return this; -998 } -999 -1000 /** -1001 * @return True if this Scan is in "raw" mode. -1002 */ -1003 public boolean isRaw() { -1004byte[] attr = getAttribute(RAW_ATTR); -1005return attr == null ? false : Bytes.toBoolean(attr); -1006 } -1007 -1008 /** -1009 * Set whether this scan is a small scan -1010 * p -1011 * Small scan should use pread and big scan can use seek + read seek + read is fast but can cause -1012 * two problem (1) resource contention (2) cause
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/071f974b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html index 54a1767..ff18b00 100644 --- a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html +++ b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.HBaseHandler.html @@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab"; -public static class ThriftServerRunner.HBaseHandler +public static class ThriftServerRunner.HBaseHandler extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface The HBaseHandler is a glue object that connects Thrift RPC calls to the @@ -721,7 +721,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface conf -protectedorg.apache.hadoop.conf.Configuration conf +protectedorg.apache.hadoop.conf.Configuration conf @@ -730,7 +730,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface LOG -protected static finalorg.apache.commons.logging.Log LOG +protected static finalorg.apache.commons.logging.Log LOG @@ -739,7 +739,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface nextScannerId -protectedint nextScannerId +protectedint nextScannerId @@ -748,7 +748,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface scannerMap -protectedhttp://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html?is-external=true; title="class or interface in java.util">HashMaphttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer,ThriftServerRunner.ResultScannerWrapper scannerMap +protectedhttp://docs.oracle.com/javase/8/docs/api/java/util/HashMap.html?is-external=true; title="class or interface in java.util">HashMaphttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer,ThriftServerRunner.ResultScannerWrapper scannerMap @@ -757,7 +757,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface metrics -privateThriftMetrics metrics +privateThriftMetrics metrics @@ -766,7 +766,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface connectionCache -private finalConnectionCache connectionCache +private finalConnectionCache connectionCache @@ -775,7 +775,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface coalescer -IncrementCoalescer coalescer +IncrementCoalescer coalescer @@ -784,7 +784,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface CLEANUP_INTERVAL -static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String CLEANUP_INTERVAL +static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String CLEANUP_INTERVAL See Also: Constant Field Values @@ -797,7 +797,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface MAX_IDLETIME -static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String MAX_IDLETIME +static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String MAX_IDLETIME See Also: Constant Field Values @@ -818,7 +818,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface HBaseHandler -protectedHBaseHandler(org.apache.hadoop.conf.Configurationc, +protectedHBaseHandler(org.apache.hadoop.conf.Configurationc, UserProvideruserProvider) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException @@ -841,7 +841,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface getAllColumns -byte[][]getAllColumns(Tabletable) +byte[][]getAllColumns(Tabletable) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Returns a list of all the column families for a given Table. @@ -858,7 +858,7 @@ implements org.apache.hadoop.hbase.thrift.generated.Hbase.Iface getTable -publicTablegetTable(byte[]tableName) +publicTablegetTable(byte[]tableName) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dc4e5c85/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html index 7c59e27..c904c56 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.DeleteColumnFamilyFuture.html @@ -119,4048 +119,4054 @@ 111import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService; 112import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest; 113import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearRegionBlockCacheRequest; -114import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest; -115import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest; -116import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest; -117import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse; -118import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest; -119import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse; -120import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest; -121import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest; -122import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos; -123import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest; -124import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse; -125import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos; -126import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription; -127import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType; -128import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema; -129import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos; -130import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest; -131import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse; -132import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest; -133import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse; -134import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest; -135import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest; -136import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest; -137import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse; -138import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest; -139import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse; -140import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest; -141import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse; -142import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest; -143import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse; -144import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest; -145import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest; -146import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse; -147import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest; -148import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse; -149import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest; -150import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse; -151import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest; -152import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse; -153import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest; -154import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest; -155import
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4abd958d/devapidocs/org/apache/hadoop/hbase/HConstants.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/HConstants.html b/devapidocs/org/apache/hadoop/hbase/HConstants.html index 8bf6d22..2340381 100644 --- a/devapidocs/org/apache/hadoop/hbase/HConstants.html +++ b/devapidocs/org/apache/hadoop/hbase/HConstants.html @@ -623,279 +623,283 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String -DEFAULT_TEMPORARY_HDFS_DIRECTORY +DEFAULT_STATUS_MULTICAST_PUBLISHER_BIND_ADDRESS +static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String +DEFAULT_TEMPORARY_HDFS_DIRECTORY + + static int DEFAULT_THREAD_WAKE_FREQUENCY Default value for thread wake frequency - + static boolean DEFAULT_USE_META_REPLICAS - + static int DEFAULT_VERSION_FILE_WRITE_ATTEMPTS Parameter name for how often we should try to write a version file, before failing - + static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String DEFAULT_WAL_STORAGE_POLICY - + static int DEFAULT_ZK_SESSION_TIMEOUT Default value for ZooKeeper session timeout - + static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String DEFAULT_ZOOKEEPER_ZNODE_PARENT - + static int DEFAULT_ZOOKEPER_CLIENT_PORT Default client port that the zookeeper listens on - + static int DEFAULT_ZOOKEPER_MAX_CLIENT_CNXNS Default limit on concurrent client-side zookeeper connections - + static int DELIMITER delimiter used between portions of a region name - + static byte[] EMPTY_BYTE_ARRAY An empty instance. - + static http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true; title="class or interface in java.nio">ByteBuffer EMPTY_BYTE_BUFFER - + static byte[] EMPTY_END_ROW Last row in a table. - + static byte[] EMPTY_START_ROW Used by scanners, etc when they want to start at the beginning of a region - + static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String ENABLE_CLIENT_BACKPRESSURE Config key for if the server should send backpressure and if the client should listen to that backpressure from the server - + static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String ENABLE_DATA_FILE_UMASK Enable file permission modification from standard hbase - + static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String ENABLE_WAL_COMPRESSION Configuration name of WAL Compression - + static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String ENABLE_WAL_ENCRYPTION Configuration key for enabling WAL encryption, a boolean - + static TableName ENSEMBLE_TABLE_NAME The name of the ensemble table - + static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String FILE_SYSTEM_VERSION Current version of file system. - + static int FOREVER Unlimited time-to-live. - + static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String HBASE_BALANCER_MAX_BALANCING Config for the max balancing time - + static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String HBASE_BALANCER_PERIOD Config for the balancer period - + static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String HBASE_CANARY_READ_RAW_SCAN_KEY - + static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String HBASE_CANARY_WRITE_DATA_TTL_KEY Canary config keys - + static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String HBASE_CANARY_WRITE_PERSERVER_REGIONS_LOWERLIMIT_KEY - + static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String HBASE_CANARY_WRITE_PERSERVER_REGIONS_UPPERLIMIT_KEY - + static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e23b49ba/devapidocs/org/apache/hadoop/hbase/HColumnDescriptor.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/HColumnDescriptor.html b/devapidocs/org/apache/hadoop/hbase/HColumnDescriptor.html index ed18c40..eb9af9e 100644 --- a/devapidocs/org/apache/hadoop/hbase/HColumnDescriptor.html +++ b/devapidocs/org/apache/hadoop/hbase/HColumnDescriptor.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":42,"i1":42,"i2":42,"i3":42,"i4":42,"i5":42,"i6":42,"i7":42,"i8":42,"i9":42,"i10":42,"i11":41,"i12":42,"i13":42,"i14":42,"i15":42,"i16":42,"i17":42,"i18":42,"i19":42,"i20":42,"i21":42,"i22":42,"i23":42,"i24":42,"i25":42,"i26":42,"i27":41,"i28":42,"i29":42,"i30":42,"i31":42,"i32":42,"i33":42,"i34":42,"i35":42,"i36":42,"i37":42,"i38":42,"i39":42,"i40":42,"i41":41,"i42":42,"i43":42,"i44":42,"i45":41,"i46":42,"i47":42,"i48":42,"i49":42,"i50":42,"i51":42,"i52":42,"i53":42,"i54":42,"i55":42,"i56":42,"i57":42,"i58":42,"i59":42,"i60":42,"i61":42,"i62":42,"i63":42,"i64":42,"i65":42,"i66":42,"i67":42,"i68":42,"i69":42,"i70":42,"i71":42,"i72":42,"i73":42,"i74":42,"i75":42,"i76":42,"i77":42,"i78":42,"i79":42,"i80":42,"i81":42,"i82":42,"i83":42}; +var methods = {"i0":42,"i1":42,"i2":42,"i3":42,"i4":42,"i5":42,"i6":42,"i7":42,"i8":42,"i9":42,"i10":42,"i11":41,"i12":42,"i13":42,"i14":42,"i15":42,"i16":42,"i17":42,"i18":42,"i19":42,"i20":42,"i21":42,"i22":42,"i23":42,"i24":42,"i25":42,"i26":42,"i27":41,"i28":42,"i29":42,"i30":42,"i31":42,"i32":42,"i33":42,"i34":42,"i35":42,"i36":42,"i37":42,"i38":42,"i39":42,"i40":41,"i41":42,"i42":42,"i43":42,"i44":41,"i45":42,"i46":42,"i47":42,"i48":42,"i49":42,"i50":42,"i51":42,"i52":42,"i53":42,"i54":42,"i55":42,"i56":42,"i57":42,"i58":42,"i59":42,"i60":42,"i61":42,"i62":42,"i63":42,"i64":42,"i65":42,"i66":42,"i67":42,"i68":42,"i69":42,"i70":42,"i71":42,"i72":42,"i73":42,"i74":42,"i75":42,"i76":42,"i77":42,"i78":42,"i79":42,"i80":42,"i81":42,"i82":42}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"],32:["t6","Deprecated Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -792,41 +792,35 @@ implements boolean -isCacheDataInL1() -Deprecated. - - - -boolean isCacheDataOnWrite() Deprecated. - + boolean isCacheIndexesOnWrite() Deprecated. - + boolean isCompressTags() Deprecated. - + boolean isEvictBlocksOnClose() Deprecated. - + boolean isInMemory() Deprecated. - + static byte[] isLegalFamilyName(byte[]b) Deprecated. @@ -834,109 +828,113 @@ implements + boolean isMobEnabled() Deprecated. Gets whether the mob is enabled for the family. - + boolean isNewVersionBehavior() Deprecated. By default, HBase only consider timestamp in versions. - + boolean isPrefetchBlocksOnOpen() Deprecated. - + static HColumnDescriptor parseFrom(byte[]bytes) Deprecated. - + void remove(byte[]key) Deprecated. - + void removeConfiguration(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringkey) Deprecated. Remove a configuration setting represented by the key. - + HColumnDescriptor setBlockCacheEnabled(booleanvalue) Deprecated. - + HColumnDescriptor setBlocksize(intvalue) Deprecated. - + HColumnDescriptor setBloomFilterType(BloomTypevalue) Deprecated. - + HColumnDescriptor setCacheBloomsOnWrite(booleanvalue) Deprecated. - + HColumnDescriptor setCacheDataInL1(booleanvalue) -Deprecated. - +Deprecated. +Since 2.0 and will be removed in 3.0 with out any replacement. Caching data in on + heap Cache, when there are both on heap LRU Cache and Bucket Cache will no longer + be supported from 2.0. + + - + HColumnDescriptor setCacheDataOnWrite(booleanvalue) Deprecated. - + HColumnDescriptor setCacheIndexesOnWrite(booleanvalue) Deprecated. - + HColumnDescriptor setCompactionCompressionType(Compression.Algorithmvalue) Deprecated. Compression types supported in hbase. - + HColumnDescriptor setCompressionType(Compression.Algorithmvalue) Deprecated. Compression types supported in hbase. - + HColumnDescriptor setCompressTags(booleanvalue) Deprecated. Set whether the tags should be compressed along with DataBlockEncoding. - + HColumnDescriptor setConfiguration(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringkey, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringvalue) @@ -944,143 +942,143 @@ implements Setter for storing a configuration setting. - + HColumnDescriptor setDataBlockEncoding(DataBlockEncodingvalue) Deprecated. Set data block encoding algorithm used
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/devapidocs/src-html/org/apache/hadoop/hbase/client/Mutation.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/Mutation.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/Mutation.html index 00db109..9c710d6 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/Mutation.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/Mutation.html @@ -36,32 +36,32 @@ 028import java.util.NavigableMap; 029import java.util.TreeMap; 030import java.util.UUID; -031 -032import org.apache.hadoop.hbase.Cell; -033import org.apache.hadoop.hbase.CellScannable; -034import org.apache.hadoop.hbase.CellScanner; -035import org.apache.hadoop.hbase.CellUtil; -036import org.apache.hadoop.hbase.HConstants; -037import org.apache.hadoop.hbase.KeyValue; -038import org.apache.hadoop.hbase.PrivateCellUtil; -039import org.apache.hadoop.hbase.Tag; -040import org.apache.hadoop.hbase.exceptions.DeserializationException; -041import org.apache.hadoop.hbase.io.HeapSize; -042import org.apache.hadoop.hbase.protobuf.ProtobufUtil; -043import org.apache.hadoop.hbase.protobuf.generated.ClientProtos; -044import org.apache.hadoop.hbase.security.access.AccessControlConstants; -045import org.apache.hadoop.hbase.security.access.AccessControlUtil; -046import org.apache.hadoop.hbase.security.access.Permission; -047import org.apache.hadoop.hbase.security.visibility.CellVisibility; -048import org.apache.hadoop.hbase.security.visibility.VisibilityConstants; -049import org.apache.hadoop.hbase.shaded.com.google.common.collect.ArrayListMultimap; -050import org.apache.hadoop.hbase.shaded.com.google.common.collect.ListMultimap; -051import org.apache.hadoop.hbase.shaded.com.google.common.io.ByteArrayDataInput; -052import org.apache.hadoop.hbase.shaded.com.google.common.io.ByteArrayDataOutput; -053import org.apache.hadoop.hbase.shaded.com.google.common.io.ByteStreams; -054import org.apache.hadoop.hbase.util.Bytes; -055import org.apache.hadoop.hbase.util.ClassSize; -056import org.apache.yetus.audience.InterfaceAudience; +031import org.apache.hadoop.hbase.Cell; +032import org.apache.hadoop.hbase.CellScannable; +033import org.apache.hadoop.hbase.CellScanner; +034import org.apache.hadoop.hbase.CellUtil; +035import org.apache.hadoop.hbase.HConstants; +036import org.apache.hadoop.hbase.KeyValue; +037import org.apache.hadoop.hbase.PrivateCellUtil; +038import org.apache.hadoop.hbase.Tag; +039import org.apache.hadoop.hbase.exceptions.DeserializationException; +040import org.apache.hadoop.hbase.io.HeapSize; +041import org.apache.hadoop.hbase.protobuf.ProtobufUtil; +042import org.apache.hadoop.hbase.protobuf.generated.ClientProtos; +043import org.apache.hadoop.hbase.security.access.AccessControlConstants; +044import org.apache.hadoop.hbase.security.access.AccessControlUtil; +045import org.apache.hadoop.hbase.security.access.Permission; +046import org.apache.hadoop.hbase.security.visibility.CellVisibility; +047import org.apache.hadoop.hbase.security.visibility.VisibilityConstants; +048import org.apache.hadoop.hbase.util.Bytes; +049import org.apache.hadoop.hbase.util.ClassSize; +050import org.apache.yetus.audience.InterfaceAudience; +051 +052import org.apache.hadoop.hbase.shaded.com.google.common.collect.ArrayListMultimap; +053import org.apache.hadoop.hbase.shaded.com.google.common.collect.ListMultimap; +054import org.apache.hadoop.hbase.shaded.com.google.common.io.ByteArrayDataInput; +055import org.apache.hadoop.hbase.shaded.com.google.common.io.ByteArrayDataOutput; +056import org.apache.hadoop.hbase.shaded.com.google.common.io.ByteStreams; 057 058@InterfaceAudience.Public 059public abstract class Mutation extends OperationWithAttributes implements Row, CellScannable, @@ -223,360 +223,508 @@ 215if (getTTL() != Long.MAX_VALUE) { 216 map.put("ttl", getTTL()); 217} -218return map; -219 } -220 -221 private static MapString, Object cellToStringMap(Cell c) { -222MapString, Object stringMap = new HashMap(); -223stringMap.put("qualifier", Bytes.toStringBinary(c.getQualifierArray(), c.getQualifierOffset(), -224 c.getQualifierLength())); -225stringMap.put("timestamp", c.getTimestamp()); -226stringMap.put("vlen", c.getValueLength()); -227ListTag tags = PrivateCellUtil.getTags(c); -228if (tags != null) { -229 ListString tagsString = new ArrayList(tags.size()); -230 for (Tag t : tags) { -231tagsString -232.add((t.getType()) + ":" + Bytes.toStringBinary(Tag.cloneValue(t))); -233 } -234 stringMap.put("tag", tagsString); -235} -236return stringMap; -237 } -238 -239 /** -240 * Set the durability for this mutation -241 * @param d -242 */ -243 public Mutation setDurability(Durability d) { -244this.durability = d; -245return this; -246 }
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c54c242b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html index 3edfbef..9707b2c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html @@ -2459,5936 +2459,5935 @@ 2451 } 2452 2453 for (HStore s : storesToFlush) { -2454MemStoreSize flushableSize = s.getFlushableSize(); -2455 totalSizeOfFlushableStores.incMemStoreSize(flushableSize); -2456 storeFlushCtxs.put(s.getColumnFamilyDescriptor().getName(), -2457 s.createFlushContext(flushOpSeqId, tracker)); -2458// for writing stores to WAL -2459 committedFiles.put(s.getColumnFamilyDescriptor().getName(), null); -2460 storeFlushableSize.put(s.getColumnFamilyDescriptor().getName(), flushableSize); -2461 } -2462 -2463 // write the snapshot start to WAL -2464 if (wal != null !writestate.readOnly) { -2465FlushDescriptor desc = ProtobufUtil.toFlushDescriptor(FlushAction.START_FLUSH, -2466getRegionInfo(), flushOpSeqId, committedFiles); -2467// No sync. Sync is below where no updates lock and we do FlushAction.COMMIT_FLUSH -2468WALUtil.writeFlushMarker(wal, this.getReplicationScope(), getRegionInfo(), desc, false, -2469mvcc); -2470 } -2471 -2472 // Prepare flush (take a snapshot) -2473 for (StoreFlushContext flush : storeFlushCtxs.values()) { -2474flush.prepare(); -2475 } -2476} catch (IOException ex) { -2477 doAbortFlushToWAL(wal, flushOpSeqId, committedFiles); -2478 throw ex; -2479} finally { -2480 this.updatesLock.writeLock().unlock(); -2481} -2482String s = "Finished memstore snapshotting " + this + ", syncing WAL and waiting on mvcc, " + -2483"flushsize=" + totalSizeOfFlushableStores; -2484status.setStatus(s); -2485doSyncOfUnflushedWALChanges(wal, getRegionInfo()); -2486return new PrepareFlushResult(storeFlushCtxs, committedFiles, storeFlushableSize, startTime, -2487flushOpSeqId, flushedSeqId, totalSizeOfFlushableStores); -2488 } -2489 -2490 /** -2491 * Utility method broken out of internalPrepareFlushCache so that method is smaller. -2492 */ -2493 private void logFatLineOnFlush(CollectionHStore storesToFlush, long sequenceId) { -2494if (!LOG.isInfoEnabled()) { -2495 return; -2496} -2497// Log a fat line detailing what is being flushed. -2498StringBuilder perCfExtras = null; -2499if (!isAllFamilies(storesToFlush)) { -2500 perCfExtras = new StringBuilder(); -2501 for (HStore store: storesToFlush) { -2502perCfExtras.append("; ").append(store.getColumnFamilyName()); -2503perCfExtras.append("=") -2504 .append(StringUtils.byteDesc(store.getFlushableSize().getDataSize())); -2505 } -2506} -2507LOG.info("Flushing " + + storesToFlush.size() + "/" + stores.size() + -2508" column families, memstore=" + StringUtils.byteDesc(this.memstoreDataSize.get()) + -2509((perCfExtras != null perCfExtras.length() 0)? perCfExtras.toString(): "") + -2510((wal != null) ? "" : "; WAL is null, using passed sequenceid=" + sequenceId)); -2511 } -2512 -2513 private void doAbortFlushToWAL(final WAL wal, final long flushOpSeqId, -2514 final Mapbyte[], ListPath committedFiles) { -2515if (wal == null) return; -2516try { -2517 FlushDescriptor desc = ProtobufUtil.toFlushDescriptor(FlushAction.ABORT_FLUSH, -2518 getRegionInfo(), flushOpSeqId, committedFiles); -2519 WALUtil.writeFlushMarker(wal, this.getReplicationScope(), getRegionInfo(), desc, false, -2520 mvcc); -2521} catch (Throwable t) { -2522 LOG.warn("Received unexpected exception trying to write ABORT_FLUSH marker to WAL:" + -2523 StringUtils.stringifyException(t)); -2524 // ignore this since we will be aborting the RS with DSE. -2525} -2526// we have called wal.startCacheFlush(), now we have to abort it -2527 wal.abortCacheFlush(this.getRegionInfo().getEncodedNameAsBytes()); -2528 } -2529 -2530 /** -2531 * Sync unflushed WAL changes. See HBASE-8208 for details -2532 */ -2533 private static void doSyncOfUnflushedWALChanges(final WAL wal, final RegionInfo hri) -2534 throws IOException { -2535if (wal == null) { -2536 return; -2537} -2538try { -2539 wal.sync(); // ensure that flush marker is sync'ed -2540} catch (IOException ioe) { -2541 wal.abortCacheFlush(hri.getEncodedNameAsBytes()); -2542 throw
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713d773f/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html b/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html index e571d5c..3803d9a 100644 --- a/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html @@ -206,12 +206,12 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.quotas.ThrottlingException.Type org.apache.hadoop.hbase.quotas.OperationQuota.OperationType -org.apache.hadoop.hbase.quotas.SpaceViolationPolicy -org.apache.hadoop.hbase.quotas.QuotaType org.apache.hadoop.hbase.quotas.QuotaScope org.apache.hadoop.hbase.quotas.ThrottleType +org.apache.hadoop.hbase.quotas.QuotaType +org.apache.hadoop.hbase.quotas.SpaceViolationPolicy +org.apache.hadoop.hbase.quotas.ThrottlingException.Type http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713d773f/devapidocs/org/apache/hadoop/hbase/regionserver/handler/RegionReplicaFlushHandler.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/handler/RegionReplicaFlushHandler.html b/devapidocs/org/apache/hadoop/hbase/regionserver/handler/RegionReplicaFlushHandler.html index b091edd..ca8621b 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/handler/RegionReplicaFlushHandler.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/handler/RegionReplicaFlushHandler.html @@ -400,7 +400,7 @@ extends triggerFlushInPrimaryRegion -voidtriggerFlushInPrimaryRegion(HRegionregion) +voidtriggerFlushInPrimaryRegion(HRegionregion) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException, http://docs.oracle.com/javase/8/docs/api/java/lang/RuntimeException.html?is-external=true; title="class or interface in java.lang">RuntimeException http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713d773f/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html index c99a865..85b1f93 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html @@ -699,19 +699,19 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage -org.apache.hadoop.hbase.regionserver.BloomType -org.apache.hadoop.hbase.regionserver.Region.Operation org.apache.hadoop.hbase.regionserver.HRegion.FlushResult.Result -org.apache.hadoop.hbase.regionserver.ScannerContext.NextState -org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection -org.apache.hadoop.hbase.regionserver.FlushType -org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType org.apache.hadoop.hbase.regionserver.ScanType +org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection org.apache.hadoop.hbase.regionserver.TimeRangeTracker.Type +org.apache.hadoop.hbase.regionserver.ScannerContext.NextState +org.apache.hadoop.hbase.regionserver.BloomType +org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage +org.apache.hadoop.hbase.regionserver.Region.Operation +org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope +org.apache.hadoop.hbase.regionserver.FlushType org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status org.apache.hadoop.hbase.regionserver.MemStoreCompactionStrategy.Action -org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope +org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fd365a2b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTableImpl.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTableImpl.html b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTableImpl.html index 91bfe40..3ddf027 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTableImpl.html +++ b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTableImpl.html @@ -279,7 +279,7 @@ implements Thttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureT batch(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends Rowactions) -Method that does a batch call on Deletes, Gets, Puts, Increments and Appends. +Method that does a batch call on Deletes, Gets, Puts, Increments, Appends and RowMutations. @@ -1165,15 +1165,15 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureTbatch(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends Rowactions) Description copied from interface:AsyncTable -Method that does a batch call on Deletes, Gets, Puts, Increments and Appends. The ordering of - execution of the actions is not defined. Meaning if you do a Put and a Get in the same - AsyncTable.batch(java.util.List? extends org.apache.hadoop.hbase.client.Row) call, you will not necessarily be guaranteed that the Get returns what the Put - had put. +Method that does a batch call on Deletes, Gets, Puts, Increments, Appends and RowMutations. The + ordering of execution of the actions is not defined. Meaning if you do a Put and a Get in the + same AsyncTable.batch(java.util.List? extends org.apache.hadoop.hbase.client.Row) call, you will not necessarily be guaranteed that the Get returns what the + Put had put. Specified by: batchin interfaceAsyncTableAdvancedScanResultConsumer Parameters: -actions - list of Get, Put, Delete, Increment, Append objects +actions - list of Get, Put, Delete, Increment, Append, and RowMutations objects Returns: A list of http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutures that represent the result for each action. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fd365a2b/devapidocs/org/apache/hadoop/hbase/client/RegionInfo.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/RegionInfo.html b/devapidocs/org/apache/hadoop/hbase/client/RegionInfo.html index 495b3bf..66c3237 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/RegionInfo.html +++ b/devapidocs/org/apache/hadoop/hbase/client/RegionInfo.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":17,"i1":6,"i2":6,"i3":17,"i4":17,"i5":17,"i6":17,"i7":17,"i8":17,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":17,"i18":17,"i19":6,"i20":17,"i21":6,"i22":17,"i23":17,"i24":17,"i25":6,"i26":6,"i27":6,"i28":6,"i29":17,"i30":17,"i31":17,"i32":17,"i33":17,"i34":17,"i35":17,"i36":17,"i37":17,"i38":17,"i39":17}; +var methods = {"i0":17,"i1":6,"i2":6,"i3":17,"i4":17,"i5":17,"i6":17,"i7":17,"i8":17,"i9":17,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":17,"i19":17,"i20":6,"i21":17,"i22":6,"i23":17,"i24":17,"i25":17,"i26":6,"i27":6,"i28":6,"i29":6,"i30":17,"i31":17,"i32":17,"i33":17,"i34":17,"i35":17,"i36":17,"i37":17,"i38":17,"i39":17,"i40":17}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],16:["t5","Default Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -212,6 +212,12 @@ public interface containsRow(byte[]row) +static RegionInfo +createMobRegionInfo(TableNametableName) +Creates a RegionInfo object for MOB data. + + + static byte[] createRegionName(TableNametableName, byte[]startKey, @@ -220,7 +226,7 @@ public interface Make a region name of passed parameters. - + static byte[] createRegionName(TableNametableName, byte[]startKey, @@ -230,7 +236,7 @@ public interface Make a region name of passed parameters. - + static byte[] createRegionName(TableNametableName, byte[]startKey, @@ -239,7 +245,7
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b9722a17/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html index 860d344..9fa61e7 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html @@ -1604,71 +1604,72 @@ 1596}); 1597 } 1598 -1599 public void postClearDeadServers() throws IOException { -1600 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { -1601 @Override -1602 public void call(MasterObserver observer) throws IOException { -1603 observer.postClearDeadServers(this); -1604 } -1605}); -1606 } -1607 -1608 public void preDecommissionRegionServers(ListServerName servers, boolean offload) throws IOException { -1609 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { -1610 @Override -1611 public void call(MasterObserver observer) throws IOException { -1612 observer.preDecommissionRegionServers(this, servers, offload); -1613 } -1614}); -1615 } -1616 -1617 public void postDecommissionRegionServers(ListServerName servers, boolean offload) throws IOException { -1618 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { -1619 @Override -1620 public void call(MasterObserver observer) throws IOException { -1621 observer.postDecommissionRegionServers(this, servers, offload); -1622 } -1623}); -1624 } -1625 -1626 public void preListDecommissionedRegionServers() throws IOException { -1627 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { -1628 @Override -1629 public void call(MasterObserver observer) throws IOException { -1630 observer.preListDecommissionedRegionServers(this); -1631 } -1632}); -1633 } -1634 -1635 public void postListDecommissionedRegionServers() throws IOException { -1636 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { -1637 @Override -1638 public void call(MasterObserver observer) throws IOException { -1639 observer.postListDecommissionedRegionServers(this); -1640 } -1641}); -1642 } -1643 -1644 public void preRecommissionRegionServer(ServerName server, Listbyte[] encodedRegionNames) -1645 throws IOException { -1646 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { -1647 @Override -1648 public void call(MasterObserver observer) throws IOException { -1649 observer.preRecommissionRegionServer(this, server, encodedRegionNames); -1650 } -1651}); -1652 } -1653 -1654 public void postRecommissionRegionServer(ServerName server, Listbyte[] encodedRegionNames) -1655 throws IOException { -1656 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { -1657 @Override -1658 public void call(MasterObserver observer) throws IOException { -1659 observer.postRecommissionRegionServer(this, server, encodedRegionNames); -1660 } -1661}); -1662 } -1663} +1599 public void postClearDeadServers(ListServerName servers, +1600 ListServerName notClearedServers) throws IOException { +1601 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { +1602 @Override +1603 public void call(MasterObserver observer) throws IOException { +1604 observer.postClearDeadServers(this, servers, notClearedServers); +1605 } +1606}); +1607 } +1608 +1609 public void preDecommissionRegionServers(ListServerName servers, boolean offload) throws IOException { +1610 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { +1611 @Override +1612 public void call(MasterObserver observer) throws IOException { +1613 observer.preDecommissionRegionServers(this, servers, offload); +1614 } +1615}); +1616 } +1617 +1618 public void postDecommissionRegionServers(ListServerName servers, boolean offload) throws IOException { +1619 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { +1620 @Override +1621 public void call(MasterObserver observer) throws IOException { +1622 observer.postDecommissionRegionServers(this, servers, offload); +1623 } +1624}); +1625 } +1626 +1627 public void preListDecommissionedRegionServers() throws IOException { +1628
1 2 3 4 >