[23/51] [partial] hbase-site git commit: Published site at 620d70d6186fb800299bcc62ad7179fccfd1be41.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html index 68302bf..a5a8905 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html @@ -2197,1768 +2197,1775 @@ 2189 warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null); 2190} 2191 -2192for (ColumnFamilyDescriptor hcd : htd.getColumnFamilies()) { -2193 if (hcd.getTimeToLive() = 0) { -2194String message = "TTL for column family " + hcd.getNameAsString() + " must be positive."; -2195 warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null); -2196 } -2197 -2198 // check blockSize -2199 if (hcd.getBlocksize() 1024 || hcd.getBlocksize() 16 * 1024 * 1024) { -2200String message = "Block size for column family " + hcd.getNameAsString() -2201+ " must be between 1K and 16MB."; +2192// check that we have minimum 1 region replicas +2193int regionReplicas = htd.getRegionReplication(); +2194if (regionReplicas 1) { +2195 String message = "Table region replication should be at least one."; +2196 warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null); +2197} +2198 +2199for (ColumnFamilyDescriptor hcd : htd.getColumnFamilies()) { +2200 if (hcd.getTimeToLive() = 0) { +2201String message = "TTL for column family " + hcd.getNameAsString() + " must be positive."; 2202 warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null); 2203 } 2204 -2205 // check versions -2206 if (hcd.getMinVersions() 0) { -2207String message = "Min versions for column family " + hcd.getNameAsString() -2208 + " must be positive."; +2205 // check blockSize +2206 if (hcd.getBlocksize() 1024 || hcd.getBlocksize() 16 * 1024 * 1024) { +2207String message = "Block size for column family " + hcd.getNameAsString() +2208+ " must be between 1K and 16MB."; 2209 warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null); 2210 } -2211 // max versions already being checked -2212 -2213 // HBASE-13776 Setting illegal versions for ColumnFamilyDescriptor -2214 // does not throw IllegalArgumentException -2215 // check minVersions = maxVerions -2216 if (hcd.getMinVersions() hcd.getMaxVersions()) { -2217String message = "Min versions for column family " + hcd.getNameAsString() -2218+ " must be less than the Max versions."; -2219 warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null); -2220 } -2221 - // check replication scope -2223 checkReplicationScope(hcd); -2224 // check bloom filter type -2225 checkBloomFilterType(hcd); -2226 -2227 // check data replication factor, it can be 0(default value) when user has not explicitly -2228 // set the value, in this case we use default replication factor set in the file system. -2229 if (hcd.getDFSReplication() 0) { -2230String message = "HFile Replication for column family " + hcd.getNameAsString() -2231+ " must be greater than zero."; -2232 warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null); -2233 } -2234 -2235 // TODO: should we check coprocessors and encryption ? -2236} -2237 } -2238 -2239 private void checkReplicationScope(ColumnFamilyDescriptor hcd) throws IOException{ -2240// check replication scope -2241WALProtos.ScopeType scop = WALProtos.ScopeType.valueOf(hcd.getScope()); -2242if (scop == null) { -2243 String message = "Replication scope for column family " -2244 + hcd.getNameAsString() + " is " + hcd.getScope() + " which is invalid."; +2211 +2212 // check versions +2213 if (hcd.getMinVersions() 0) { +2214String message = "Min versions for column family " + hcd.getNameAsString() +2215 + " must be positive."; +2216 warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null); +2217 } +2218 // max versions already being checked +2219 +2220 // HBASE-13776 Setting illegal versions for ColumnFamilyDescriptor +2221 // does not throw IllegalArgumentException + // check minVersions = maxVerions +2223 if (hcd.getMinVersions() hcd.getMaxVersions()) { +2224String message = "Min versions for column family " + hcd.getNameAsString() +2225+ " must be less than the Max versions."; +2226 warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message,
[23/51] [partial] hbase-site git commit: Published site at 281d6429e55149cc4c05430dcc1d1dc136d8b245.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/901d593a/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html b/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html index 60a5382..5e9a1d4 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html +++ b/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html @@ -247,7 +247,7 @@ extends Methods inherited from classorg.apache.hadoop.hbase.master.HMaster -abort, abortProcedure, addColumn, addReplicationPeer, balance, balance, balanceSwitch, canCreateBaseZNode, canUpdateTableDescriptor, checkIfShouldMoveSystemRegionAsync, checkInitialized, checkServiceStarted, checkTableModifiable, configureInfoServer, constructMaster, createAssignmentManager, createMetaBootstrap, createNamespace, createRpcServices, createServerManager, createSystemTable, createTable, decommissionRegionServers, decorateMasterConfiguration, deleteColumn, deleteNamespace, deleteTable, disableReplicationPeer, disableTable, enableReplicationPeer, enableTable, getAssignmentManager, getAverageLoad, getCatalogJanitor, getClientIdAuditPrefix, getClusterMetrics, getClusterMetrics, getClusterMetricsWithoutCoprocessor, getClusterMetricsWithoutCoprocessor, getClusterSchema, getDumpServlet, getFavoredNodesManager, getHFileCleaner, getInitializedEvent, getLastMajorCompactionTimestamp, getLastMajorCompactionTimestampForRegion, getLoadBalancer, getLoadBalancerClassName, getLoadedCoprocessors, getLockManager, getLocks, getLogCleaner, getMasterActiveTime, getMasterCoprocessorHost, getMasterCoprocessors, getMasterFileSystem, getMasterFinishedInitializationTime, getMasterMetrics, getMasterProcedureExecutor, getMasterProcedureManagerHost, getMasterQuotaManager, getMasterRpcServices, getMasterStartTime, getMasterWalManager, getMergePlanCount, getMetaTableObserver, getMobCompactionState, getNamespace, getNamespaces, getNumWALFiles, getProcedures, getProcessName, getQuotaObserverChore, getRegionNormalizer, getRegionNormalizerTracker, getRegionServerFatalLogBuffer, getRegionServerInfoPort, getRegionServerVersion, getRemoteInetAddress, getReplicationLoad, getReplicationPeerConfig, getReplicationPeerManager, getServerManager, getServerName, getSnapshotM anager, getSnapshotQuotaObserverChore, getSpaceQuotaSnapshotNotifier, getSplitOrMergeTracker, getSplitPlanCount, getSyncReplicationReplayWALManager, getTableStateManager, getUseThisHostnameInstead, getWalProcedureStore, getZooKeeper, initClusterSchemaService, initializeZKBasedSystemTrackers, isActiveMaster, isBalancerOn, isCatalogJanitorEnabled, isCleanerChoreEnabled, isInitialized, isInMaintenanceMode, isNormalizerOn, isSplitOrMergeEnabled, listDecommissionedRegionServers, listReplicationPeers, listTableDescriptors, listTableDescriptorsByNamespace, listTableNames, listTableNamesByNamespace, login, main, mergeRegions, modifyColumn, modifyNamespace, modifyTable, move, normalizeRegions, recommissionRegionServer, registerService, remoteProcedureCompleted, remoteProcedureFailed, removeReplicationPeer, reportMobCompactionEnd, reportMobCompactionStart, requestMobCompaction, restoreSnapshot, setCatalogJanitorEnabled, setInitialized, shutdown, splitRegion, stop, stopMaster, stopServiceThreads, transitReplicationPeerSyncReplicationState, truncateTable, updateConfigurationForQuotasObserver, updateReplicationPeerConfig, waitForMasterActive, waitForMetaOnline +abort, abortProcedure, addColumn, addReplicationPeer, balance, balance, balanceSwitch, canCreateBaseZNode, canUpdateTableDescriptor, checkIfShouldMoveSystemRegionAsync, checkInitialized, checkServiceStarted, checkTableModifiable, configureInfoServer, constructMaster, createAssignmentManager, createMetaBootstrap, createNamespace, createRpcServices, createServerManager, createSystemTable, createTable, decommissionRegionServers, decorateMasterConfiguration, deleteColumn, deleteNamespace, deleteTable, disableReplicationPeer, disableTable, enableReplicationPeer, enableTable, getAssignmentManager, getAverageLoad, getCatalogJanitor, getClientIdAuditPrefix, getClusterMetrics, getClusterMetrics, getClusterMetricsWithoutCoprocessor, getClusterMetricsWithoutCoprocessor, getClusterSchema, getDumpServlet, getFavoredNodesManager, getHFileCleaner, getInitializedEvent, getLastMajorCompactionTimestamp, getLastMajorCompactionTimestampForRegion, getLoadBalancer, getLoadBalancerClassName, getLoadedCoprocessors, getLockManager, getLocks, getLogCleaner, getMasterActiveTime, getMasterCoprocessorHost, getMasterCoprocessors, getMasterFileSystem, getMasterFinishedInitializationTime,
[23/51] [partial] hbase-site git commit: Published site at 466fa920fee572fe20db3b77ebf539dc304d5f31.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2bf59208/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncNonMetaRegionLocator.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncNonMetaRegionLocator.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncNonMetaRegionLocator.html index 9aa9b59..ac7e0ea 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncNonMetaRegionLocator.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncNonMetaRegionLocator.html @@ -28,520 +28,565 @@ 020import static org.apache.hadoop.hbase.HConstants.NINES; 021import static org.apache.hadoop.hbase.HConstants.ZEROES; 022import static org.apache.hadoop.hbase.TableName.META_TABLE_NAME; -023import static org.apache.hadoop.hbase.client.ConnectionUtils.createClosestRowAfter; -024import static org.apache.hadoop.hbase.client.ConnectionUtils.isEmptyStopRow; -025import static org.apache.hadoop.hbase.client.RegionInfo.createRegionName; -026import static org.apache.hadoop.hbase.util.Bytes.BYTES_COMPARATOR; -027import static org.apache.hadoop.hbase.util.CollectionUtils.computeIfAbsent; -028 -029import java.io.IOException; -030import java.util.Arrays; -031import java.util.HashSet; -032import java.util.Iterator; -033import java.util.LinkedHashMap; -034import java.util.Map; -035import java.util.Optional; -036import java.util.Set; -037import java.util.concurrent.CompletableFuture; -038import java.util.concurrent.ConcurrentHashMap; -039import java.util.concurrent.ConcurrentMap; -040import java.util.concurrent.ConcurrentNavigableMap; -041import java.util.concurrent.ConcurrentSkipListMap; -042import org.apache.hadoop.hbase.DoNotRetryIOException; -043import org.apache.hadoop.hbase.HConstants; -044import org.apache.hadoop.hbase.HRegionLocation; -045import org.apache.hadoop.hbase.MetaTableAccessor; -046import org.apache.hadoop.hbase.RegionLocations; -047import org.apache.hadoop.hbase.TableName; -048import org.apache.hadoop.hbase.TableNotFoundException; -049import org.apache.hadoop.hbase.client.Scan.ReadType; -050import org.apache.hadoop.hbase.util.Bytes; -051import org.apache.yetus.audience.InterfaceAudience; -052import org.slf4j.Logger; -053import org.slf4j.LoggerFactory; -054 -055import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; -056 -057/** -058 * The asynchronous locator for regions other than meta. -059 */ -060@InterfaceAudience.Private -061class AsyncNonMetaRegionLocator { -062 -063 private static final Logger LOG = LoggerFactory.getLogger(AsyncNonMetaRegionLocator.class); +023import static org.apache.hadoop.hbase.client.AsyncRegionLocatorHelper.canUpdateOnError; +024import static org.apache.hadoop.hbase.client.AsyncRegionLocatorHelper.createRegionLocations; +025import static org.apache.hadoop.hbase.client.AsyncRegionLocatorHelper.isGood; +026import static org.apache.hadoop.hbase.client.AsyncRegionLocatorHelper.mergeRegionLocations; +027import static org.apache.hadoop.hbase.client.AsyncRegionLocatorHelper.removeRegionLocation; +028import static org.apache.hadoop.hbase.client.ConnectionUtils.createClosestRowAfter; +029import static org.apache.hadoop.hbase.client.ConnectionUtils.isEmptyStopRow; +030import static org.apache.hadoop.hbase.client.RegionInfo.createRegionName; +031import static org.apache.hadoop.hbase.util.Bytes.BYTES_COMPARATOR; +032import static org.apache.hadoop.hbase.util.CollectionUtils.computeIfAbsent; +033 +034import java.io.IOException; +035import java.util.Arrays; +036import java.util.HashSet; +037import java.util.Iterator; +038import java.util.LinkedHashMap; +039import java.util.Map; +040import java.util.Optional; +041import java.util.Set; +042import java.util.concurrent.CompletableFuture; +043import java.util.concurrent.ConcurrentHashMap; +044import java.util.concurrent.ConcurrentMap; +045import java.util.concurrent.ConcurrentNavigableMap; +046import java.util.concurrent.ConcurrentSkipListMap; +047import org.apache.commons.lang3.ObjectUtils; +048import org.apache.hadoop.hbase.DoNotRetryIOException; +049import org.apache.hadoop.hbase.HBaseIOException; +050import org.apache.hadoop.hbase.HConstants; +051import org.apache.hadoop.hbase.HRegionLocation; +052import org.apache.hadoop.hbase.MetaTableAccessor; +053import org.apache.hadoop.hbase.RegionLocations; +054import org.apache.hadoop.hbase.TableName; +055import org.apache.hadoop.hbase.TableNotFoundException; +056import org.apache.hadoop.hbase.client.Scan.ReadType; +057import org.apache.hadoop.hbase.util.Bytes; +058import org.apache.yetus.audience.InterfaceAudience; +059import org.slf4j.Logger; +060import org.slf4j.LoggerFactory; +061 +062import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; +063import org.apache.hbase.thirdparty.com.google.common.base.Objects; 064 -065 @VisibleForTesting -066 static final
[23/51] [partial] hbase-site git commit: Published site at e4b6b4afb933a961f543537875f87a2dc62d3757.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/devapidocs/src-html/org/apache/hadoop/hbase/thrift/Constants.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/Constants.html b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/Constants.html new file mode 100644 index 000..d28d77e --- /dev/null +++ b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/Constants.html @@ -0,0 +1,223 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + +Source code + + + + +001/* +002 * Licensed to the Apache Software Foundation (ASF) under one +003 * or more contributor license agreements. See the NOTICE file +004 * distributed with this work for additional information +005 * regarding copyright ownership. The ASF licenses this file +006 * to you under the Apache License, Version 2.0 (the +007 * "License"); you may not use this file except in compliance +008 * with the License. You may obtain a copy of the License at +009 * +010 * http://www.apache.org/licenses/LICENSE-2.0 +011 * +012 * Unless required by applicable law or agreed to in writing, software +013 * distributed under the License is distributed on an "AS IS" BASIS, +014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +015 * See the License for the specific language governing permissions and +016 * limitations under the License. +017 */ +018package org.apache.hadoop.hbase.thrift; +019 +020import org.apache.yetus.audience.InterfaceAudience; +021 +022/** +023 * Thrift related constants +024 */ +025@InterfaceAudience.Private +026public final class Constants { +027 private Constants(){} +028 +029 public static final int DEFAULT_HTTP_MAX_HEADER_SIZE = 64 * 1024; // 64k +030 +031 public static final String SERVER_TYPE_CONF_KEY = +032 "hbase.regionserver.thrift.server.type"; +033 +034 public static final String COMPACT_CONF_KEY = "hbase.regionserver.thrift.compact"; +035 public static final boolean COMPACT_CONF_DEFAULT = false; +036 +037 public static final String FRAMED_CONF_KEY = "hbase.regionserver.thrift.framed"; +038 public static final boolean FRAMED_CONF_DEFAULT = false; +039 +040 public static final String MAX_FRAME_SIZE_CONF_KEY = +041 "hbase.regionserver.thrift.framed.max_frame_size_in_mb"; +042 public static final int MAX_FRAME_SIZE_CONF_DEFAULT = 2; +043 +044 public static final String COALESCE_INC_KEY = "hbase.regionserver.thrift.coalesceIncrement"; +045 public static final String USE_HTTP_CONF_KEY = "hbase.regionserver.thrift.http"; +046 +047 public static final String HTTP_MIN_THREADS_KEY = "hbase.thrift.http_threads.min"; +048 public static final int HTTP_MIN_THREADS_KEY_DEFAULT = 2; +049 +050 public static final String HTTP_MAX_THREADS_KEY = "hbase.thrift.http_threads.max"; +051 public static final int HTTP_MAX_THREADS_KEY_DEFAULT = 100; +052 +053 // ssl related configs +054 public static final String THRIFT_SSL_ENABLED_KEY = "hbase.thrift.ssl.enabled"; +055 public static final String THRIFT_SSL_KEYSTORE_STORE_KEY = "hbase.thrift.ssl.keystore.store"; +056 public static final String THRIFT_SSL_KEYSTORE_PASSWORD_KEY = +057 "hbase.thrift.ssl.keystore.password"; +058 public static final String THRIFT_SSL_KEYSTORE_KEYPASSWORD_KEY +059 = "hbase.thrift.ssl.keystore.keypassword"; +060 public static final String THRIFT_SSL_EXCLUDE_CIPHER_SUITES_KEY = +061 "hbase.thrift.ssl.exclude.cipher.suites"; +062 public static final String THRIFT_SSL_INCLUDE_CIPHER_SUITES_KEY = +063 "hbase.thrift.ssl.include.cipher.suites"; +064 public static final String THRIFT_SSL_EXCLUDE_PROTOCOLS_KEY = +065 "hbase.thrift.ssl.exclude.protocols"; +066 public static final String THRIFT_SSL_INCLUDE_PROTOCOLS_KEY = +067 "hbase.thrift.ssl.include.protocols"; +068 +069 +070 public static final String THRIFT_SUPPORT_PROXYUSER_KEY = "hbase.thrift.support.proxyuser"; +071 +072 //kerberos related configs +073 public static final String THRIFT_DNS_INTERFACE_KEY = "hbase.thrift.dns.interface"; +074 public static final String THRIFT_DNS_NAMESERVER_KEY = "hbase.thrift.dns.nameserver"; +075 public static final String THRIFT_KERBEROS_PRINCIPAL_KEY = "hbase.thrift.kerberos.principal"; +076 public static final String THRIFT_KEYTAB_FILE_KEY = "hbase.thrift.keytab.file"; +077 public static final String THRIFT_SPNEGO_PRINCIPAL_KEY = "hbase.thrift.spnego.principal"; +078 public static final String THRIFT_SPNEGO_KEYTAB_FILE_KEY = "hbase.thrift.spnego.keytab.file"; +079 +080 /** +081 * Amount of time in milliseconds before a server thread will timeout +082 * waiting for client to send data on a connected socket. Currently, +083 * applies only to TBoundedThreadPoolServer +084 */ +085 public static final String THRIFT_SERVER_SOCKET_READ_TIMEOUT_KEY = +086 "hbase.thrift.server.socket.read.timeout"; +087 public static final int
[23/51] [partial] hbase-site git commit: Published site at 3ab895979b643a2980bcdb7fee2078f14b614210.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/codec/package-summary.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/codec/package-summary.html b/devapidocs/org/apache/hadoop/hbase/codec/package-summary.html index 14dcf5b..16c4959 100644 --- a/devapidocs/org/apache/hadoop/hbase/codec/package-summary.html +++ b/devapidocs/org/apache/hadoop/hbase/codec/package-summary.html @@ -275,6 +275,6 @@ -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/codec/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/codec/package-tree.html b/devapidocs/org/apache/hadoop/hbase/codec/package-tree.html index 4468eff..eb81cc7 100644 --- a/devapidocs/org/apache/hadoop/hbase/codec/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/codec/package-tree.html @@ -204,6 +204,6 @@ -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/codec/package-use.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/codec/package-use.html b/devapidocs/org/apache/hadoop/hbase/codec/package-use.html index 76343f5..41324f1 100644 --- a/devapidocs/org/apache/hadoop/hbase/codec/package-use.html +++ b/devapidocs/org/apache/hadoop/hbase/codec/package-use.html @@ -277,6 +277,6 @@ -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationManager.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationManager.html b/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationManager.html index 6ea799e..34baa82 100644 --- a/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationManager.html +++ b/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationManager.html @@ -420,6 +420,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationObserver.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationObserver.html b/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationObserver.html index 72b04d4..d588c0a 100644 --- a/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationObserver.html +++ b/devapidocs/org/apache/hadoop/hbase/conf/ConfigurationObserver.html @@ -235,6 +235,6 @@ public interface Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/conf/PropagatingConfigurationObserver.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/conf/PropagatingConfigurationObserver.html b/devapidocs/org/apache/hadoop/hbase/conf/PropagatingConfigurationObserver.html index 880a869..6896ab9 100644 --- a/devapidocs/org/apache/hadoop/hbase/conf/PropagatingConfigurationObserver.html +++ b/devapidocs/org/apache/hadoop/hbase/conf/PropagatingConfigurationObserver.html @@ -265,6 +265,6 @@ extends Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/conf/class-use/ConfigurationManager.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/conf/class-use/ConfigurationManager.html b/devapidocs/org/apache/hadoop/hbase/conf/class-use/ConfigurationManager.html index 143db46..960e9f4 100644 ---
[23/51] [partial] hbase-site git commit: Published site at 7820ba1dbdba58b1002cdfde08eb21aa7a0bb6da.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html index 4ade4c1..03e1383 100644 --- a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html @@ -199,9 +199,9 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) +org.apache.hadoop.hbase.thrift.MetricsThriftServerSourceFactoryImpl.FactoryStorage org.apache.hadoop.hbase.thrift.ThriftMetrics.ThriftServerType org.apache.hadoop.hbase.thrift.ThriftServerRunner.ImplType -org.apache.hadoop.hbase.thrift.MetricsThriftServerSourceFactoryImpl.FactoryStorage http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.THBaseServiceMetricsProxy.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.THBaseServiceMetricsProxy.html b/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.THBaseServiceMetricsProxy.html index 8b8ba09..a21f2ad 100644 --- a/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.THBaseServiceMetricsProxy.html +++ b/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.THBaseServiceMetricsProxy.html @@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab"; -private static final class ThriftHBaseServiceHandler.THBaseServiceMetricsProxy +private static final class ThriftHBaseServiceHandler.THBaseServiceMetricsProxy extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements https://docs.oracle.com/javase/8/docs/api/java/lang/reflect/InvocationHandler.html?is-external=true; title="class or interface in java.lang.reflect">InvocationHandler @@ -215,7 +215,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/lang/reflect/ handler -private finalorg.apache.hadoop.hbase.thrift2.generated.THBaseService.Iface handler +private finalorg.apache.hadoop.hbase.thrift2.generated.THBaseService.Iface handler @@ -224,7 +224,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/lang/reflect/ metrics -private finalThriftMetrics metrics +private finalThriftMetrics metrics @@ -241,7 +241,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/lang/reflect/ THBaseServiceMetricsProxy -privateTHBaseServiceMetricsProxy(org.apache.hadoop.hbase.thrift2.generated.THBaseService.Ifacehandler, +privateTHBaseServiceMetricsProxy(org.apache.hadoop.hbase.thrift2.generated.THBaseService.Ifacehandler, ThriftMetricsmetrics) @@ -259,7 +259,7 @@ implements https://docs.oracle.com/javase/8/docs/api/java/lang/reflect/ invoke -publichttps://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Objectinvoke(https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Objectproxy, +publichttps://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Objectinvoke(https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Objectproxy, https://docs.oracle.com/javase/8/docs/api/java/lang/reflect/Method.html?is-external=true; title="class or interface in java.lang.reflect">Methodm, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object[]args) throws https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwable http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.TIOErrorWithCause.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.TIOErrorWithCause.html b/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.TIOErrorWithCause.html index 24184b7..bf9ac4e 100644 ---
[23/51] [partial] hbase-site git commit: Published site at c448604ceb987d113913f0583452b2abce04db0d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html index 84c4071..e7b6097 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.LogDelegate.html @@ -105,7 +105,7 @@ var activeTableTab = "activeTableTab"; -static interface RSRpcServices.LogDelegate +static interface RSRpcServices.LogDelegate @@ -151,7 +151,7 @@ var activeTableTab = "activeTableTab"; logBatchWarning -voidlogBatchWarning(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringfirstRegionName, +voidlogBatchWarning(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringfirstRegionName, intsum, introwSizeWarnThreshold) http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html index 5d0e72a..9e736dc 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerCloseCallBack.html @@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab"; -private static final class RSRpcServices.RegionScannerCloseCallBack +private static final class RSRpcServices.RegionScannerCloseCallBack extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements RpcCallback An Rpc callback for closing a RegionScanner. @@ -209,7 +209,7 @@ implements scanner -private finalRegionScanner scanner +private finalRegionScanner scanner @@ -226,7 +226,7 @@ implements RegionScannerCloseCallBack -publicRegionScannerCloseCallBack(RegionScannerscanner) +publicRegionScannerCloseCallBack(RegionScannerscanner) @@ -243,7 +243,7 @@ implements run -publicvoidrun() +publicvoidrun() throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Description copied from interface:RpcCallback Called at the end of an Rpc Call RpcCallContext http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html index 70b0bd0..69d484d 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.RegionScannerHolder.html @@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab"; -private static final class RSRpcServices.RegionScannerHolder +private static final class RSRpcServices.RegionScannerHolder extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object Holder class which holds the RegionScanner, nextCallSeq and RpcCallbacks together. @@ -239,7 +239,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html nextCallSeq -private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicLong.html?is-external=true; title="class or interface in java.util.concurrent.atomic">AtomicLong nextCallSeq +private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicLong.html?is-external=true; title="class or interface in java.util.concurrent.atomic">AtomicLong nextCallSeq @@ -248,7 +248,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html scannerName -private finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String scannerName +private finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String scannerName @@ -257,7 +257,7 @@ extends
[23/51] [partial] hbase-site git commit: Published site at 8bf966c8e936dec4d83bcbe85c5aab543f14a0df.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/27555316/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.html index ea05301..26a93dd 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.html @@ -269,3590 +269,3574 @@ 261 */ 262 protected ClusterConnection clusterConnection; 263 -264 /* -265 * Long-living meta table locator, which is created when the server is started and stopped -266 * when server shuts down. References to this locator shall be used to perform according -267 * operations in EventHandlers. Primary reason for this decision is to make it mockable -268 * for tests. -269 */ -270 protected MetaTableLocator metaTableLocator; -271 -272 /** -273 * Go here to get table descriptors. -274 */ -275 protected TableDescriptors tableDescriptors; -276 -277 // Replication services. If no replication, this handler will be null. -278 protected ReplicationSourceService replicationSourceHandler; -279 protected ReplicationSinkService replicationSinkHandler; -280 -281 // Compactions -282 public CompactSplit compactSplitThread; -283 -284 /** -285 * Map of regions currently being served by this region server. Key is the -286 * encoded region name. All access should be synchronized. -287 */ -288 protected final MapString, HRegion onlineRegions = new ConcurrentHashMap(); -289 -290 /** -291 * Map of encoded region names to the DataNode locations they should be hosted on -292 * We store the value as InetSocketAddress since this is used only in HDFS -293 * API (create() that takes favored nodes as hints for placing file blocks). -294 * We could have used ServerName here as the value class, but we'd need to -295 * convert it to InetSocketAddress at some point before the HDFS API call, and -296 * it seems a bit weird to store ServerName since ServerName refers to RegionServers -297 * and here we really mean DataNode locations. -298 */ -299 protected final MapString, InetSocketAddress[] regionFavoredNodesMap = -300 new ConcurrentHashMap(); -301 -302 // Leases -303 protected Leases leases; +264 /** +265 * Go here to get table descriptors. +266 */ +267 protected TableDescriptors tableDescriptors; +268 +269 // Replication services. If no replication, this handler will be null. +270 protected ReplicationSourceService replicationSourceHandler; +271 protected ReplicationSinkService replicationSinkHandler; +272 +273 // Compactions +274 public CompactSplit compactSplitThread; +275 +276 /** +277 * Map of regions currently being served by this region server. Key is the +278 * encoded region name. All access should be synchronized. +279 */ +280 protected final MapString, HRegion onlineRegions = new ConcurrentHashMap(); +281 +282 /** +283 * Map of encoded region names to the DataNode locations they should be hosted on +284 * We store the value as InetSocketAddress since this is used only in HDFS +285 * API (create() that takes favored nodes as hints for placing file blocks). +286 * We could have used ServerName here as the value class, but we'd need to +287 * convert it to InetSocketAddress at some point before the HDFS API call, and +288 * it seems a bit weird to store ServerName since ServerName refers to RegionServers +289 * and here we really mean DataNode locations. +290 */ +291 protected final MapString, InetSocketAddress[] regionFavoredNodesMap = +292 new ConcurrentHashMap(); +293 +294 // Leases +295 protected Leases leases; +296 +297 // Instance of the hbase executor executorService. +298 protected ExecutorService executorService; +299 +300 // If false, the file system has become unavailable +301 protected volatile boolean fsOk; +302 protected HFileSystem fs; +303 protected HFileSystem walFs; 304 -305 // Instance of the hbase executor executorService. -306 protected ExecutorService executorService; -307 -308 // If false, the file system has become unavailable -309 protected volatile boolean fsOk; -310 protected HFileSystem fs; -311 protected HFileSystem walFs; -312 -313 // Set when a report to the master comes back with a message asking us to -314 // shutdown. Also set by call to stop when debugging or running unit tests -315 // of HRegionServer in isolation. -316 private volatile boolean stopped = false; -317 -318 // Go down hard. Used if file system becomes unavailable and also in -319 // debugging and unit tests. -320 private volatile boolean abortRequested; -321 public static final String ABORT_TIMEOUT = "hbase.regionserver.abort.timeout"; -322 // Default abort timeout
[23/51] [partial] hbase-site git commit: Published site at 1acbd36c903b048141866b143507bfce124a5c5f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyNamespaceProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyNamespaceProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyNamespaceProcedure.html index cb0eb2d..a08c4ed 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyNamespaceProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyNamespaceProcedure.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -129,7 +129,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public class ModifyNamespaceProcedure +public class ModifyNamespaceProcedure extends AbstractStateMachineNamespaceProcedureorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyNamespaceState The procedure to add a namespace to an existing table. @@ -191,10 +191,6 @@ extends private NamespaceDescriptor oldNsDescriptor - -private https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean -traceEnabled - @@ -210,6 +206,13 @@ extends Procedure NO_PROC_ID, NO_TIMEOUT + + + + +Fields inherited from interfaceorg.apache.hadoop.hbase.master.procedure.TableProcedureInterface +DUMMY_NAMESPACE_TABLE_NAME + @@ -287,67 +290,45 @@ extends -private TableNamespaceManager -getTableNamespaceManager(MasterProcedureEnvenv) - - TableProcedureInterface.TableOperationType getTableOperationType() Given an operation type we can take decisions about what to do with pending operations. - -private void -insertIntoNSTable(MasterProcedureEnvenv) -Insert/update the row into namespace table - - - + protected boolean isRollbackSupported(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyNamespaceStatestate) Used by the default implementation of abort() to know if the current state can be aborted and rollback can be triggered. - -private https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean -isTraceEnabled() -The procedure could be restarted from a different machine. - - - + private boolean prepareModify(MasterProcedureEnvenv) Action before any real action of adding namespace. - + protected void rollbackState(MasterProcedureEnvenv, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyNamespaceStatestate) called to perform the rollback of the specified state - + protected void serializeStateData(ProcedureStateSerializerserializer) The user-level code of the procedure may have some state to persist (e.g. - -private void -updateZKNamespaceManager(MasterProcedureEnvenv) -Update ZooKeeper. - - Methods inherited from classorg.apache.hadoop.hbase.master.procedure.AbstractStateMachineNamespaceProcedure -acquireLock, getTableName, releaseLock, releaseSyncLatch, toStringClassDetails, waitInitialized +acquireLock, addOrUpdateNamespace, createDirectory, createDirectory, getTableName, getTableNamespaceManager, releaseLock, releaseSyncLatch, toStringClassDetails, waitInitialized @@ -390,7 +371,7 @@ extends LOG -private static finalorg.slf4j.Logger LOG +private static finalorg.slf4j.Logger LOG @@ -399,25 +380,16 @@ extends oldNsDescriptor -privateNamespaceDescriptor oldNsDescriptor +privateNamespaceDescriptor oldNsDescriptor - - -newNsDescriptor -privateNamespaceDescriptor newNsDescriptor - - - - - -traceEnabled -privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean traceEnabled +newNsDescriptor +privateNamespaceDescriptor newNsDescriptor @@ -434,7 +406,7 @@ extends ModifyNamespaceProcedure -publicModifyNamespaceProcedure() +publicModifyNamespaceProcedure() @@ -443,7 +415,7 @@ extends ModifyNamespaceProcedure -publicModifyNamespaceProcedure(MasterProcedureEnvenv, +publicModifyNamespaceProcedure(MasterProcedureEnvenv, NamespaceDescriptornewNsDescriptor) @@ -453,7 +425,7 @@ extends ModifyNamespaceProcedure -publicModifyNamespaceProcedure(MasterProcedureEnvenv, +publicModifyNamespaceProcedure(MasterProcedureEnvenv, NamespaceDescriptornewNsDescriptor,
[23/51] [partial] hbase-site git commit: Published site at 130057f13774f6b213cdb06952c805a29d59396e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/68eae623/devapidocs/src-html/org/apache/hadoop/hbase/security/access/AccessController.OpType.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/security/access/AccessController.OpType.html b/devapidocs/src-html/org/apache/hadoop/hbase/security/access/AccessController.OpType.html index 25b7848..5c428b5 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/security/access/AccessController.OpType.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/security/access/AccessController.OpType.html @@ -245,20 +245,20 @@ 237return regionEnv != null ? regionEnv.getRegion() : null; 238 } 239 -240 public TableAuthManager getAuthManager() { +240 public AuthManager getAuthManager() { 241return accessChecker.getAuthManager(); 242 } 243 244 private void initialize(RegionCoprocessorEnvironment e) throws IOException { 245final Region region = e.getRegion(); 246Configuration conf = e.getConfiguration(); -247Mapbyte[], ListMultimapString, TablePermission tables = AccessControlLists.loadAll(region); +247Mapbyte[], ListMultimapString, UserPermission tables = AccessControlLists.loadAll(region); 248// For each table, write out the table's permissions to the respective 249// znode for that table. -250for (Map.Entrybyte[], ListMultimapString,TablePermission t: +250for (Map.Entrybyte[], ListMultimapString, UserPermission t: 251 tables.entrySet()) { 252 byte[] entry = t.getKey(); -253 ListMultimapString,TablePermission perms = t.getValue(); +253 ListMultimapString, UserPermission perms = t.getValue(); 254 byte[] serialized = AccessControlLists.writePermissionsAsBytes(perms, conf); 255 getAuthManager().getZKPermissionWatcher().writeToZookeeper(entry, serialized); 256} @@ -294,7 +294,7 @@ 286try (Table t = e.getConnection().getTable(AccessControlLists.ACL_TABLE_NAME)) { 287 for (byte[] entry : entries) { 288currentEntry = entry; -289ListMultimapString, TablePermission perms = +289ListMultimapString, UserPermission perms = 290 AccessControlLists.getPermissions(conf, entry, t, null, null, null, false); 291byte[] serialized = AccessControlLists.writePermissionsAsBytes(perms, conf); 292zkw.writeToZookeeper(entry, serialized); @@ -338,7 +338,7 @@ 330} 331 332// 2. check for the table-level, if successful we can short-circuit -333if (getAuthManager().authorize(user, tableName, (byte[])null, permRequest)) { +333if (getAuthManager().authorizeUserTable(user, tableName, permRequest)) { 334 return AuthResult.allow(request, "Table permission granted", user, 335permRequest, tableName, families); 336} @@ -348,7 +348,7 @@ 340 // all families must pass 341 for (Map.Entrybyte [], ? extends Collection? family : families.entrySet()) { 342// a) check for family level access -343if (getAuthManager().authorize(user, tableName, family.getKey(), +343if (getAuthManager().authorizeUserTable(user, tableName, family.getKey(), 344permRequest)) { 345 continue; // family-level permission overrides per-qualifier 346} @@ -359,17 +359,17 @@ 351// for each qualifier of the family 352Setbyte[] familySet = (Setbyte[])family.getValue(); 353for (byte[] qualifier : familySet) { -354 if (!getAuthManager().authorize(user, tableName, family.getKey(), -355 qualifier, permRequest)) { +354 if (!getAuthManager().authorizeUserTable(user, tableName, +355family.getKey(), qualifier, permRequest)) { 356return AuthResult.deny(request, "Failed qualifier check", user, -357permRequest, tableName, makeFamilyMap(family.getKey(), qualifier)); +357 permRequest, tableName, makeFamilyMap(family.getKey(), qualifier)); 358 } 359} 360 } else if (family.getValue() instanceof List) { // ListCell 361ListCell cellList = (ListCell)family.getValue(); 362for (Cell cell : cellList) { -363 if (!getAuthManager().authorize(user, tableName, family.getKey(), -364 CellUtil.cloneQualifier(cell), permRequest)) { +363 if (!getAuthManager().authorizeUserTable(user, tableName, family.getKey(), +364 CellUtil.cloneQualifier(cell), permRequest)) { 365return AuthResult.deny(request, "Failed qualifier check", user, permRequest, 366 tableName, makeFamilyMap(family.getKey(), CellUtil.cloneQualifier(cell))); 367 } @@ -378,7 +378,7 @@ 370} else { 371 // no qualifiers and family-level check already failed
[23/51] [partial] hbase-site git commit: Published site at 3fe8649b2c9ba1271c25e8f476548907e4c7a90d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8f09a71d/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.ReadRequestCostFunction.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.ReadRequestCostFunction.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.ReadRequestCostFunction.html index c7d99b2..9d1542c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.ReadRequestCostFunction.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.ReadRequestCostFunction.html @@ -382,1357 +382,1365 @@ 374for (int i = 0; i this.curFunctionCosts.length; i++) { 375 curFunctionCosts[i] = tempFunctionCosts[i]; 376} -377LOG.info("start StochasticLoadBalancer.balancer, initCost=" + currentCost + ", functionCost=" -378+ functionCost()); +377double initCost = currentCost; +378double newCost = currentCost; 379 -380double initCost = currentCost; -381double newCost = currentCost; -382 -383long computedMaxSteps; -384if (runMaxSteps) { -385 computedMaxSteps = Math.max(this.maxSteps, -386 ((long)cluster.numRegions * (long)this.stepsPerRegion * (long)cluster.numServers)); -387} else { -388 computedMaxSteps = Math.min(this.maxSteps, -389 ((long)cluster.numRegions * (long)this.stepsPerRegion * (long)cluster.numServers)); -390} -391// Perform a stochastic walk to see if we can get a good fit. -392long step; -393 -394for (step = 0; step computedMaxSteps; step++) { -395 Cluster.Action action = nextAction(cluster); -396 -397 if (action.type == Type.NULL) { -398continue; -399 } -400 -401 cluster.doAction(action); -402 updateCostsWithAction(cluster, action); -403 -404 newCost = computeCost(cluster, currentCost); -405 -406 // Should this be kept? -407 if (newCost currentCost) { -408currentCost = newCost; -409 -410// save for JMX -411curOverallCost = currentCost; -412for (int i = 0; i this.curFunctionCosts.length; i++) { -413 curFunctionCosts[i] = tempFunctionCosts[i]; -414} -415 } else { -416// Put things back the way they were before. -417// TODO: undo by remembering old values -418Action undoAction = action.undoAction(); -419cluster.doAction(undoAction); -420updateCostsWithAction(cluster, undoAction); -421 } -422 -423 if (EnvironmentEdgeManager.currentTime() - startTime -424 maxRunningTime) { -425break; -426 } -427} -428long endTime = EnvironmentEdgeManager.currentTime(); -429 -430 metricsBalancer.balanceCluster(endTime - startTime); -431 -432// update costs metrics -433updateStochasticCosts(tableName, curOverallCost, curFunctionCosts); -434if (initCost currentCost) { -435 plans = createRegionPlans(cluster); -436 LOG.info("Finished computing new load balance plan. Computation took {}" + -437" to try {} different iterations. Found a solution that moves " + -438"{} regions; Going from a computed cost of {}" + -439" to a new cost of {}", java.time.Duration.ofMillis(endTime - startTime), -440step, plans.size(), initCost, currentCost); -441 return plans; -442} -443LOG.info("Could not find a better load balance plan. Tried {} different configurations in " + -444 "{}, and did not find anything with a computed cost less than {}", step, -445 java.time.Duration.ofMillis(endTime - startTime), initCost); -446return null; -447 } -448 -449 /** -450 * update costs to JMX -451 */ -452 private void updateStochasticCosts(TableName tableName, Double overall, Double[] subCosts) { -453if (tableName == null) return; -454 -455// check if the metricsBalancer is MetricsStochasticBalancer before casting -456if (metricsBalancer instanceof MetricsStochasticBalancer) { -457 MetricsStochasticBalancer balancer = (MetricsStochasticBalancer) metricsBalancer; -458 // overall cost -459 balancer.updateStochasticCost(tableName.getNameAsString(), -460"Overall", "Overall cost", overall); -461 -462 // each cost function -463 for (int i = 0; i costFunctions.length; i++) { -464CostFunction costFunction = costFunctions[i]; -465String costFunctionName = costFunction.getClass().getSimpleName(); -466Double costPercent = (overall == 0) ? 0 : (subCosts[i] / overall); -467// TODO: cost function may need a specific description -468 balancer.updateStochasticCost(tableName.getNameAsString(), costFunctionName, -469 "The percent of " + costFunctionName, costPercent); -470 } -471} -472 }
[23/51] [partial] hbase-site git commit: Published site at 7adf590106826b9e4432cfeee06acdc0ccff8c6e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/425db230/testapidocs/src-html/org/apache/hadoop/hbase/HBaseTestingUtility.html -- diff --git a/testapidocs/src-html/org/apache/hadoop/hbase/HBaseTestingUtility.html b/testapidocs/src-html/org/apache/hadoop/hbase/HBaseTestingUtility.html index 0c894de..8729895 100644 --- a/testapidocs/src-html/org/apache/hadoop/hbase/HBaseTestingUtility.html +++ b/testapidocs/src-html/org/apache/hadoop/hbase/HBaseTestingUtility.html @@ -179,4145 +179,4146 @@ 171 * avoiding port contention if another local HBase instance is already running). 172 * pTo preserve test data directories, pass the system property "hbase.testing.preserve.testdir" 173 * setting it to true. -174 */ -175@InterfaceAudience.Public -176@SuppressWarnings("deprecation") -177public class HBaseTestingUtility extends HBaseZKTestingUtility { -178 -179 /** -180 * System property key to get test directory value. Name is as it is because mini dfs has -181 * hard-codings to put test data here. It should NOT be used directly in HBase, as it's a property -182 * used in mini dfs. -183 * @deprecated can be used only with mini dfs -184 */ -185 @Deprecated -186 private static final String TEST_DIRECTORY_KEY = "test.build.data"; -187 -188 public static final String REGIONS_PER_SERVER_KEY = "hbase.test.regions-per-server"; -189 /** -190 * The default number of regions per regionserver when creating a pre-split -191 * table. -192 */ -193 public static final int DEFAULT_REGIONS_PER_SERVER = 3; -194 +174 * Trigger pre commit. +175 */ +176@InterfaceAudience.Public +177@SuppressWarnings("deprecation") +178public class HBaseTestingUtility extends HBaseZKTestingUtility { +179 +180 /** +181 * System property key to get test directory value. Name is as it is because mini dfs has +182 * hard-codings to put test data here. It should NOT be used directly in HBase, as it's a property +183 * used in mini dfs. +184 * @deprecated can be used only with mini dfs +185 */ +186 @Deprecated +187 private static final String TEST_DIRECTORY_KEY = "test.build.data"; +188 +189 public static final String REGIONS_PER_SERVER_KEY = "hbase.test.regions-per-server"; +190 /** +191 * The default number of regions per regionserver when creating a pre-split +192 * table. +193 */ +194 public static final int DEFAULT_REGIONS_PER_SERVER = 3; 195 -196 public static final String PRESPLIT_TEST_TABLE_KEY = "hbase.test.pre-split-table"; -197 public static final boolean PRESPLIT_TEST_TABLE = true; -198 -199 private MiniDFSCluster dfsCluster = null; -200 -201 private volatile HBaseCluster hbaseCluster = null; -202 private MiniMRCluster mrCluster = null; -203 -204 /** If there is a mini cluster running for this testing utility instance. */ -205 private volatile boolean miniClusterRunning; -206 -207 private String hadoopLogDir; -208 -209 /** Directory on test filesystem where we put the data for this instance of -210* HBaseTestingUtility*/ -211 private Path dataTestDirOnTestFS = null; -212 -213 /** -214 * Shared cluster connection. -215 */ -216 private volatile Connection connection; -217 -218 /** Filesystem URI used for map-reduce mini-cluster setup */ -219 private static String FS_URI; -220 -221 /** This is for unit tests parameterized with a single boolean. */ -222 public static final ListObject[] MEMSTORETS_TAGS_PARAMETRIZED = memStoreTSAndTagsCombination(); -223 -224 /** -225 * Checks to see if a specific port is available. -226 * -227 * @param port the port number to check for availability -228 * @return tttrue/tt if the port is available, or ttfalse/tt if not -229 */ -230 public static boolean available(int port) { -231ServerSocket ss = null; -232DatagramSocket ds = null; -233try { -234 ss = new ServerSocket(port); -235 ss.setReuseAddress(true); -236 ds = new DatagramSocket(port); -237 ds.setReuseAddress(true); -238 return true; -239} catch (IOException e) { -240 // Do nothing -241} finally { -242 if (ds != null) { -243ds.close(); -244 } -245 -246 if (ss != null) { -247try { -248 ss.close(); -249} catch (IOException e) { -250 /* should not be thrown */ -251} -252 } -253} -254 -255return false; -256 } -257 -258 /** -259 * Create all combinations of Bloom filters and compression algorithms for -260 * testing. -261 */ -262 private static ListObject[] bloomAndCompressionCombinations() { -263ListObject[] configurations = new ArrayList(); -264for (Compression.Algorithm comprAlgo : -265 HBaseCommonTestingUtility.COMPRESSION_ALGORITHMS) { -266 for (BloomType bloomType : BloomType.values()) { -267configurations.add(new Object[] { comprAlgo, bloomType }); -268 } -269} -270return
[23/51] [partial] hbase-site git commit: Published site at 5fbb227deb365fe812d433fe39b85ac4b0ddee20.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureUtil.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureUtil.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureUtil.html index 2be5d7b..90d9eed 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureUtil.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureUtil.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9}; +var methods = {"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -111,7 +111,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private @InterfaceStability.Evolving -public final class MasterProcedureUtil +public final class MasterProcedureUtil extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object @@ -221,6 +221,12 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html toUserInfo(org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationuserInfoProto) +static https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException +unwrapRemoteIOException(Procedureproc) +This is a version of unwrapRemoteIOException that can do DoNotRetryIOE. + + + static boolean validateProcedureWALFilename(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringfilename) A Procedure WAL file name is of the format: pv-wal-id.log where wal-id is 20 digits. @@ -254,7 +260,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html pattern -private static finalhttps://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Pattern pattern +private static finalhttps://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Pattern pattern Pattern used to validate a Procedure WAL file name see validateProcedureWALFilename(String) for description. @@ -273,7 +279,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html MasterProcedureUtil -privateMasterProcedureUtil() +privateMasterProcedureUtil() @@ -290,7 +296,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html toProtoUserInfo -public staticorg.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationtoProtoUserInfo(Useruser) +public staticorg.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationtoProtoUserInfo(Useruser) @@ -299,7 +305,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html toUserInfo -public staticUsertoUserInfo(org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationuserInfoProto) +public staticUsertoUserInfo(org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationuserInfoProto) @@ -308,7 +314,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html submitProcedure -public staticlongsubmitProcedure(MasterProcedureUtil.NonceProcedureRunnablerunnable) +public staticlongsubmitProcedure(MasterProcedureUtil.NonceProcedureRunnablerunnable) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Helper used to deal with submitting procs with nonce. Internally the NonceProcedureRunnable.run() will be called only if no one else @@ -337,7 +343,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html validateProcedureWALFilename -public staticbooleanvalidateProcedureWALFilename(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringfilename) +public staticbooleanvalidateProcedureWALFilename(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringfilename) A Procedure WAL file name is of the format: pv-wal-id.log where wal-id is 20 digits. Parameters: @@ -353,7 +359,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html getTablePriority -public staticintgetTablePriority(TableNametableName) +public staticintgetTablePriority(TableNametableName) Return the priority for the given table. Now meta table is 3, other system tables are 2, and user tables are 1. @@ -361,14 +367,27
[23/51] [partial] hbase-site git commit: Published site at 821e4d7de2d576189f4288d1c2acf9e9a9471f5c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/323b17d9/testdevapidocs/org/apache/hadoop/hbase/client/TestMobRestoreSnapshotFromClientClone.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/client/TestMobRestoreSnapshotFromClientClone.html b/testdevapidocs/org/apache/hadoop/hbase/client/TestMobRestoreSnapshotFromClientClone.html new file mode 100644 index 000..9fd5cd8 --- /dev/null +++ b/testdevapidocs/org/apache/hadoop/hbase/client/TestMobRestoreSnapshotFromClientClone.html @@ -0,0 +1,423 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +TestMobRestoreSnapshotFromClientClone (Apache HBase 3.0.0-SNAPSHOT Test API) + + + + + +var methods = {"i0":10,"i1":10,"i2":9,"i3":9,"i4":10}; +var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Field| +Constr| +Method + + +Detail: +Field| +Constr| +Method + + + + + + + + +org.apache.hadoop.hbase.client +Class TestMobRestoreSnapshotFromClientClone + + + +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.client.RestoreSnapshotFromClientTestBase + + +org.apache.hadoop.hbase.client.RestoreSnapshotFromClientCloneTestBase + + +org.apache.hadoop.hbase.client.TestMobRestoreSnapshotFromClientClone + + + + + + + + + + + + +public class TestMobRestoreSnapshotFromClientClone +extends RestoreSnapshotFromClientCloneTestBase + + + + + + + + + + + +Field Summary + +Fields + +Modifier and Type +Field and Description + + +static HBaseClassTestRule +CLASS_RULE + + + + + + +Fields inherited from classorg.apache.hadoop.hbase.client.RestoreSnapshotFromClientTestBase +admin, emptySnapshot, FAMILY, name, snapshot0Rows, snapshot1Rows, snapshotName0, snapshotName1, snapshotName2, tableName, TEST_FAMILY2, TEST_UTIL + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +TestMobRestoreSnapshotFromClientClone() + + + + + + + + + +Method Summary + +All MethodsStatic MethodsInstance MethodsConcrete Methods + +Modifier and Type +Method and Description + + +protected int +countRows(org.apache.hadoop.hbase.client.Tabletable, + byte[]...families) + + +protected void +createTable() + + +static void +setupCluster() + + +protected static void +setupConf(org.apache.hadoop.conf.Configurationconf) + + +protected void +verifyRowCount(HBaseTestingUtilityutil, + org.apache.hadoop.hbase.TableNametableName, + longexpectedRows) + + + + + + +Methods inherited from classorg.apache.hadoop.hbase.client.RestoreSnapshotFromClientCloneTestBase +testCloneAndRestoreSnapshot, testCloneSnapshotOfCloned + + + + + +Methods inherited from classorg.apache.hadoop.hbase.client.RestoreSnapshotFromClientTestBase +getNumReplicas, getValidMethodName, setup, splitRegion, tearDown, tearDownAfterClass + + + + + +Methods inherited from classjava.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--; title="class or interface in java.lang">clone, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-; title="class or interface in java.lang">equals, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--; title="class or interface in java.lang">finalize, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--; title="class or interface in java.lang">getClass, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--; title="class or interface in java.lang">hashCode, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--; title="class or interface in java.lang">notify, https://docs.oracle.com/javase/8/docs/api/ja va/lang/Object.html?is-external=true#notifyAll--" title="class or interface in
[23/51] [partial] hbase-site git commit: Published site at fa5fa6ecdd071b72b58971058ff3ab9d28c3e709.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d1341859/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html index 061ce80..bdfc3f8 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.html @@ -39,2126 +39,2163 @@ 031import java.util.Set; 032import java.util.concurrent.ConcurrentHashMap; 033import java.util.concurrent.CopyOnWriteArrayList; -034import java.util.concurrent.TimeUnit; -035import java.util.concurrent.atomic.AtomicBoolean; -036import java.util.concurrent.atomic.AtomicInteger; -037import java.util.concurrent.atomic.AtomicLong; -038import java.util.stream.Collectors; -039import java.util.stream.Stream; -040import org.apache.hadoop.conf.Configuration; -041import org.apache.hadoop.hbase.HConstants; -042import org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException; -043import org.apache.hadoop.hbase.log.HBaseMarkers; -044import org.apache.hadoop.hbase.procedure2.Procedure.LockState; -045import org.apache.hadoop.hbase.procedure2.store.ProcedureStore; -046import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator; -047import org.apache.hadoop.hbase.procedure2.util.StringUtils; -048import org.apache.hadoop.hbase.security.User; -049import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -050import org.apache.hadoop.hbase.util.IdLock; -051import org.apache.hadoop.hbase.util.NonceKey; -052import org.apache.hadoop.hbase.util.Threads; -053import org.apache.yetus.audience.InterfaceAudience; -054import org.slf4j.Logger; -055import org.slf4j.LoggerFactory; -056 -057import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; -058import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; +034import java.util.concurrent.Executor; +035import java.util.concurrent.Executors; +036import java.util.concurrent.TimeUnit; +037import java.util.concurrent.atomic.AtomicBoolean; +038import java.util.concurrent.atomic.AtomicInteger; +039import java.util.concurrent.atomic.AtomicLong; +040import java.util.stream.Collectors; +041import java.util.stream.Stream; +042import org.apache.hadoop.conf.Configuration; +043import org.apache.hadoop.hbase.HConstants; +044import org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException; +045import org.apache.hadoop.hbase.log.HBaseMarkers; +046import org.apache.hadoop.hbase.procedure2.Procedure.LockState; +047import org.apache.hadoop.hbase.procedure2.store.ProcedureStore; +048import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator; +049import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureStoreListener; +050import org.apache.hadoop.hbase.procedure2.util.StringUtils; +051import org.apache.hadoop.hbase.security.User; +052import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; +053import org.apache.hadoop.hbase.util.IdLock; +054import org.apache.hadoop.hbase.util.NonceKey; +055import org.apache.hadoop.hbase.util.Threads; +056import org.apache.yetus.audience.InterfaceAudience; +057import org.slf4j.Logger; +058import org.slf4j.LoggerFactory; 059 -060import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState; -061 -062/** -063 * Thread Pool that executes the submitted procedures. -064 * The executor has a ProcedureStore associated. -065 * Each operation is logged and on restart the pending procedures are resumed. -066 * -067 * Unless the Procedure code throws an error (e.g. invalid user input) -068 * the procedure will complete (at some point in time), On restart the pending -069 * procedures are resumed and the once failed will be rolledback. +060import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; +061import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; +062import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; +063 +064import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState; +065 +066/** +067 * Thread Pool that executes the submitted procedures. +068 * The executor has a ProcedureStore associated. +069 * Each operation is logged and on restart the pending procedures are resumed. 070 * -071 * The user can add procedures to the executor via submitProcedure(proc) -072 * check for the finished state via isFinished(procId) -073 * and get the result via getResult(procId) -074 */ -075@InterfaceAudience.Private -076public class ProcedureExecutorTEnvironment { -077 private static final Logger LOG = LoggerFactory.getLogger(ProcedureExecutor.class); -078 -079 public static final String
[23/51] [partial] hbase-site git commit: Published site at 6bc7089f9e0793efc9bdd46a84f5ccd9bc4579ad.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/419d0338/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html index 43c66a8..061ce80 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html @@ -23,2136 +23,2142 @@ 015 * See the License for the specific language governing permissions and 016 * limitations under the License. 017 */ -018 -019package org.apache.hadoop.hbase.procedure2; -020 -021import java.io.IOException; -022import java.util.ArrayDeque; -023import java.util.ArrayList; -024import java.util.Arrays; -025import java.util.Collection; -026import java.util.Deque; -027import java.util.HashSet; -028import java.util.Iterator; -029import java.util.List; -030import java.util.Map; -031import java.util.Objects; -032import java.util.Set; -033import java.util.concurrent.ConcurrentHashMap; -034import java.util.concurrent.CopyOnWriteArrayList; -035import java.util.concurrent.TimeUnit; -036import java.util.concurrent.atomic.AtomicBoolean; -037import java.util.concurrent.atomic.AtomicInteger; -038import java.util.concurrent.atomic.AtomicLong; -039import java.util.stream.Collectors; -040import java.util.stream.Stream; -041 -042import org.apache.hadoop.conf.Configuration; -043import org.apache.hadoop.hbase.HConstants; -044import org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException; -045import org.apache.hadoop.hbase.log.HBaseMarkers; -046import org.apache.hadoop.hbase.procedure2.Procedure.LockState; -047import org.apache.hadoop.hbase.procedure2.store.ProcedureStore; -048import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator; -049import org.apache.hadoop.hbase.procedure2.util.StringUtils; -050import org.apache.hadoop.hbase.security.User; -051import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -052import org.apache.hadoop.hbase.util.IdLock; -053import org.apache.hadoop.hbase.util.NonceKey; -054import org.apache.hadoop.hbase.util.Threads; -055import org.apache.yetus.audience.InterfaceAudience; -056import org.slf4j.Logger; -057import org.slf4j.LoggerFactory; -058 -059import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; -060import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; +018package org.apache.hadoop.hbase.procedure2; +019 +020import java.io.IOException; +021import java.util.ArrayDeque; +022import java.util.ArrayList; +023import java.util.Arrays; +024import java.util.Collection; +025import java.util.Deque; +026import java.util.HashSet; +027import java.util.Iterator; +028import java.util.List; +029import java.util.Map; +030import java.util.Objects; +031import java.util.Set; +032import java.util.concurrent.ConcurrentHashMap; +033import java.util.concurrent.CopyOnWriteArrayList; +034import java.util.concurrent.TimeUnit; +035import java.util.concurrent.atomic.AtomicBoolean; +036import java.util.concurrent.atomic.AtomicInteger; +037import java.util.concurrent.atomic.AtomicLong; +038import java.util.stream.Collectors; +039import java.util.stream.Stream; +040import org.apache.hadoop.conf.Configuration; +041import org.apache.hadoop.hbase.HConstants; +042import org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException; +043import org.apache.hadoop.hbase.log.HBaseMarkers; +044import org.apache.hadoop.hbase.procedure2.Procedure.LockState; +045import org.apache.hadoop.hbase.procedure2.store.ProcedureStore; +046import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator; +047import org.apache.hadoop.hbase.procedure2.util.StringUtils; +048import org.apache.hadoop.hbase.security.User; +049import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; +050import org.apache.hadoop.hbase.util.IdLock; +051import org.apache.hadoop.hbase.util.NonceKey; +052import org.apache.hadoop.hbase.util.Threads; +053import org.apache.yetus.audience.InterfaceAudience; +054import org.slf4j.Logger; +055import org.slf4j.LoggerFactory; +056 +057import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; +058import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; +059 +060import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState; 061 -062import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState; -063 -064/** -065 * Thread Pool that executes the submitted procedures. -066 * The executor has a ProcedureStore associated. -067 * Each operation is logged and on restart the pending procedures
[23/51] [partial] hbase-site git commit: Published site at d7e08317d2f214e4cca7b67578aba0ed7a567d54.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.WriterThread.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.WriterThread.html b/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.WriterThread.html index aa153a2..5aa7403 100644 --- a/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.WriterThread.html +++ b/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.WriterThread.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -public static class WALSplitter.WriterThread +public static class WALSplitter.WriterThread extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true; title="class or interface in java.lang">Thread @@ -266,7 +266,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html shouldStop -private volatileboolean shouldStop +private volatileboolean shouldStop @@ -275,7 +275,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html controller -privateWALSplitter.PipelineController controller +privateWALSplitter.PipelineController controller @@ -284,7 +284,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html entryBuffers -privateWALSplitter.EntryBuffers entryBuffers +privateWALSplitter.EntryBuffers entryBuffers @@ -293,7 +293,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html outputSink -privateWALSplitter.OutputSink outputSink +privateWALSplitter.OutputSink outputSink @@ -310,7 +310,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html WriterThread -WriterThread(WALSplitter.PipelineControllercontroller, +WriterThread(WALSplitter.PipelineControllercontroller, WALSplitter.EntryBuffersentryBuffers, WALSplitter.OutputSinksink, inti) @@ -330,7 +330,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html run -publicvoidrun() +publicvoidrun() Specified by: https://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true#run--; title="class or interface in java.lang">runin interfacehttps://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true; title="class or interface in java.lang">Runnable @@ -345,7 +345,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html doRun -privatevoiddoRun() +privatevoiddoRun() throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Throws: @@ -359,7 +359,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html writeBuffer -privatevoidwriteBuffer(WALSplitter.RegionEntryBufferbuffer) +privatevoidwriteBuffer(WALSplitter.RegionEntryBufferbuffer) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Throws: @@ -373,7 +373,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html finish -voidfinish() +voidfinish() http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.html b/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.html index 0257e4b..621546b 100644 --- a/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.html +++ b/devapidocs/org/apache/hadoop/hbase/wal/WALSplitter.html @@ -228,75 +228,75 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html fileBeingSplit -protected org.apache.hadoop.fs.FileSystem -fs - - protected https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">Long lastFlushedSequenceIds - + private static org.slf4j.Logger LOG - + private static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String OLD_SEQUENCE_ID_FILE_SUFFIX - + (package private) WALSplitter.OutputSink outputSink - + private static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String RECOVERED_LOG_TMPFILE_SUFFIX - + protected https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in
[23/51] [partial] hbase-site git commit: Published site at 8eaaa63114a64bcaeaf0ed9bdd88615ee22255c1.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f6f9d4f3/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html index 0cf012a..976894f 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.MasterStoppedException.html @@ -63,3884 +63,3883 @@ 055import javax.servlet.http.HttpServlet; 056import javax.servlet.http.HttpServletRequest; 057import javax.servlet.http.HttpServletResponse; -058 -059import org.apache.commons.lang3.StringUtils; -060import org.apache.hadoop.conf.Configuration; -061import org.apache.hadoop.fs.Path; -062import org.apache.hadoop.hbase.ChoreService; -063import org.apache.hadoop.hbase.ClusterId; -064import org.apache.hadoop.hbase.ClusterMetrics; -065import org.apache.hadoop.hbase.ClusterMetrics.Option; -066import org.apache.hadoop.hbase.ClusterMetricsBuilder; -067import org.apache.hadoop.hbase.CompoundConfiguration; -068import org.apache.hadoop.hbase.DoNotRetryIOException; -069import org.apache.hadoop.hbase.HBaseIOException; -070import org.apache.hadoop.hbase.HBaseInterfaceAudience; -071import org.apache.hadoop.hbase.HConstants; -072import org.apache.hadoop.hbase.InvalidFamilyOperationException; -073import org.apache.hadoop.hbase.MasterNotRunningException; -074import org.apache.hadoop.hbase.MetaTableAccessor; -075import org.apache.hadoop.hbase.NamespaceDescriptor; -076import org.apache.hadoop.hbase.PleaseHoldException; -077import org.apache.hadoop.hbase.ReplicationPeerNotFoundException; -078import org.apache.hadoop.hbase.ServerName; -079import org.apache.hadoop.hbase.TableDescriptors; -080import org.apache.hadoop.hbase.TableName; -081import org.apache.hadoop.hbase.TableNotDisabledException; -082import org.apache.hadoop.hbase.TableNotFoundException; -083import org.apache.hadoop.hbase.UnknownRegionException; -084import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor; -085import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder; -086import org.apache.hadoop.hbase.client.MasterSwitchType; -087import org.apache.hadoop.hbase.client.RegionInfo; -088import org.apache.hadoop.hbase.client.RegionInfoBuilder; -089import org.apache.hadoop.hbase.client.Result; -090import org.apache.hadoop.hbase.client.TableDescriptor; -091import org.apache.hadoop.hbase.client.TableDescriptorBuilder; -092import org.apache.hadoop.hbase.client.TableState; -093import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; -094import org.apache.hadoop.hbase.exceptions.DeserializationException; -095import org.apache.hadoop.hbase.exceptions.MergeRegionException; -096import org.apache.hadoop.hbase.executor.ExecutorType; -097import org.apache.hadoop.hbase.favored.FavoredNodesManager; -098import org.apache.hadoop.hbase.favored.FavoredNodesPromoter; -099import org.apache.hadoop.hbase.http.InfoServer; -100import org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils; -101import org.apache.hadoop.hbase.ipc.RpcServer; -102import org.apache.hadoop.hbase.ipc.ServerNotRunningYetException; -103import org.apache.hadoop.hbase.log.HBaseMarkers; -104import org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode; -105import org.apache.hadoop.hbase.master.assignment.AssignProcedure; -106import org.apache.hadoop.hbase.master.assignment.AssignmentManager; -107import org.apache.hadoop.hbase.master.assignment.MergeTableRegionsProcedure; -108import org.apache.hadoop.hbase.master.assignment.MoveRegionProcedure; -109import org.apache.hadoop.hbase.master.assignment.RegionStateNode; -110import org.apache.hadoop.hbase.master.assignment.RegionStates; -111import org.apache.hadoop.hbase.master.assignment.TransitRegionStateProcedure; -112import org.apache.hadoop.hbase.master.assignment.UnassignProcedure; -113import org.apache.hadoop.hbase.master.balancer.BalancerChore; -114import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer; -115import org.apache.hadoop.hbase.master.balancer.ClusterStatusChore; -116import org.apache.hadoop.hbase.master.balancer.LoadBalancerFactory; -117import org.apache.hadoop.hbase.master.cleaner.CleanerChore; -118import org.apache.hadoop.hbase.master.cleaner.HFileCleaner; -119import org.apache.hadoop.hbase.master.cleaner.LogCleaner; -120import org.apache.hadoop.hbase.master.cleaner.ReplicationBarrierCleaner; -121import org.apache.hadoop.hbase.master.locking.LockManager; -122import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan; -123import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan.PlanType; -124import org.apache.hadoop.hbase.master.normalizer.RegionNormalizer; -125import
[23/51] [partial] hbase-site git commit: Published site at cd161d976ef47b84e904f2d54bac65d2f3417c2a.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fa1bebf8/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TableFuture.TableWaitForStateCallable.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TableFuture.TableWaitForStateCallable.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TableFuture.TableWaitForStateCallable.html index a5789e0..93a57cb 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TableFuture.TableWaitForStateCallable.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TableFuture.TableWaitForStateCallable.html @@ -238,4120 +238,4119 @@ 230 * @see Admin 231 */ 232@InterfaceAudience.Private -233@InterfaceStability.Evolving -234public class HBaseAdmin implements Admin { -235 private static final Logger LOG = LoggerFactory.getLogger(HBaseAdmin.class); -236 -237 private ClusterConnection connection; -238 -239 private final Configuration conf; -240 private final long pause; -241 private final int numRetries; -242 private final int syncWaitTimeout; -243 private boolean aborted; -244 private int operationTimeout; -245 private int rpcTimeout; -246 -247 private RpcRetryingCallerFactory rpcCallerFactory; -248 private RpcControllerFactory rpcControllerFactory; -249 -250 private NonceGenerator ng; -251 -252 @Override -253 public int getOperationTimeout() { -254return operationTimeout; -255 } -256 -257 HBaseAdmin(ClusterConnection connection) throws IOException { -258this.conf = connection.getConfiguration(); -259this.connection = connection; -260 -261// TODO: receive ConnectionConfiguration here rather than re-parsing these configs every time. -262this.pause = this.conf.getLong(HConstants.HBASE_CLIENT_PAUSE, -263 HConstants.DEFAULT_HBASE_CLIENT_PAUSE); -264this.numRetries = this.conf.getInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, -265 HConstants.DEFAULT_HBASE_CLIENT_RETRIES_NUMBER); -266this.operationTimeout = this.conf.getInt(HConstants.HBASE_CLIENT_OPERATION_TIMEOUT, -267 HConstants.DEFAULT_HBASE_CLIENT_OPERATION_TIMEOUT); -268this.rpcTimeout = this.conf.getInt(HConstants.HBASE_RPC_TIMEOUT_KEY, -269 HConstants.DEFAULT_HBASE_RPC_TIMEOUT); -270this.syncWaitTimeout = this.conf.getInt( -271 "hbase.client.sync.wait.timeout.msec", 10 * 6); // 10min -272 -273this.rpcCallerFactory = connection.getRpcRetryingCallerFactory(); -274this.rpcControllerFactory = connection.getRpcControllerFactory(); -275 -276this.ng = this.connection.getNonceGenerator(); -277 } -278 -279 @Override -280 public void abort(String why, Throwable e) { -281// Currently does nothing but throw the passed message and exception -282this.aborted = true; -283throw new RuntimeException(why, e); -284 } -285 -286 @Override -287 public boolean isAborted() { -288return this.aborted; -289 } -290 -291 @Override -292 public boolean abortProcedure(final long procId, final boolean mayInterruptIfRunning) -293 throws IOException { -294return get(abortProcedureAsync(procId, mayInterruptIfRunning), this.syncWaitTimeout, -295 TimeUnit.MILLISECONDS); -296 } -297 -298 @Override -299 public FutureBoolean abortProcedureAsync(final long procId, final boolean mayInterruptIfRunning) -300 throws IOException { -301Boolean abortProcResponse = -302executeCallable(new MasterCallableAbortProcedureResponse(getConnection(), -303getRpcControllerFactory()) { -304 @Override -305 protected AbortProcedureResponse rpcCall() throws Exception { -306AbortProcedureRequest abortProcRequest = -307 AbortProcedureRequest.newBuilder().setProcId(procId).build(); -308return master.abortProcedure(getRpcController(), abortProcRequest); -309 } -310}).getIsProcedureAborted(); -311return new AbortProcedureFuture(this, procId, abortProcResponse); -312 } -313 -314 @Override -315 public ListTableDescriptor listTableDescriptors() throws IOException { -316return listTableDescriptors((Pattern)null, false); -317 } -318 -319 @Override -320 public ListTableDescriptor listTableDescriptors(Pattern pattern) throws IOException { -321return listTableDescriptors(pattern, false); -322 } -323 -324 @Override -325 public ListTableDescriptor listTableDescriptors(Pattern pattern, boolean includeSysTables) -326 throws IOException { -327return executeCallable(new MasterCallableListTableDescriptor(getConnection(), -328getRpcControllerFactory()) { -329 @Override -330 protected ListTableDescriptor rpcCall() throws Exception { -331GetTableDescriptorsRequest req = -332 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables); -333return
[23/51] [partial] hbase-site git commit: Published site at c6a65ba63fce85ac7c4b62b96ef2bbe6c35d2f00.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/293abb17/devapidocs/src-html/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html b/devapidocs/src-html/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html index 309387e..11a1dff 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html @@ -27,211 +27,206 @@ 019 020import java.io.IOException; 021import java.util.ArrayList; -022import java.util.Arrays; -023import java.util.Comparator; -024import java.util.Objects; -025import java.util.TreeSet; -026 -027import org.apache.hadoop.hbase.Cell; -028import org.apache.hadoop.hbase.CellUtil; -029import org.apache.hadoop.hbase.PrivateCellUtil; -030import org.apache.yetus.audience.InterfaceAudience; -031import org.apache.hadoop.hbase.exceptions.DeserializationException; -032import org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException; -033import org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations; -034import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos; -035import org.apache.hadoop.hbase.util.Bytes; -036 -037/** -038 * This filter is used for selecting only those keys with columns that matches -039 * a particular prefix. For example, if prefix is 'an', it will pass keys will -040 * columns like 'and', 'anti' but not keys with columns like 'ball', 'act'. -041 */ -042@InterfaceAudience.Public -043public class MultipleColumnPrefixFilter extends FilterBase { -044 protected byte [] hint = null; -045 protected TreeSetbyte [] sortedPrefixes = createTreeSet(); -046 private final static int MAX_LOG_PREFIXES = 5; -047 -048 public MultipleColumnPrefixFilter(final byte [][] prefixes) { -049if (prefixes != null) { -050 for (int i = 0; i prefixes.length; i++) { -051if (!sortedPrefixes.add(prefixes[i])) -052 throw new IllegalArgumentException ("prefixes must be distinct"); -053 } -054} -055 } -056 -057 public byte [][] getPrefix() { -058int count = 0; -059byte [][] temp = new byte [sortedPrefixes.size()][]; -060for (byte [] prefixes : sortedPrefixes) { -061 temp [count++] = prefixes; -062} -063return temp; -064 } -065 -066 @Override -067 public boolean filterRowKey(Cell cell) throws IOException { -068// Impl in FilterBase might do unnecessary copy for Off heap backed Cells. -069return false; -070 } -071 -072 @Deprecated -073 @Override -074 public ReturnCode filterKeyValue(final Cell c) { -075return filterCell(c); -076 } -077 -078 @Override -079 public ReturnCode filterCell(final Cell c) { -080if (sortedPrefixes.isEmpty()) { -081 return ReturnCode.INCLUDE; -082} else { -083 return filterColumn(c); -084} -085 } -086 -087 public ReturnCode filterColumn(Cell cell) { -088byte [] qualifier = CellUtil.cloneQualifier(cell); -089TreeSetbyte [] lesserOrEqualPrefixes = -090 (TreeSetbyte []) sortedPrefixes.headSet(qualifier, true); -091 -092if (lesserOrEqualPrefixes.size() != 0) { -093 byte [] largestPrefixSmallerThanQualifier = lesserOrEqualPrefixes.last(); -094 -095 if (Bytes.startsWith(qualifier, largestPrefixSmallerThanQualifier)) { -096return ReturnCode.INCLUDE; -097 } -098 -099 if (lesserOrEqualPrefixes.size() == sortedPrefixes.size()) { -100return ReturnCode.NEXT_ROW; -101 } else { -102hint = sortedPrefixes.higher(largestPrefixSmallerThanQualifier); -103return ReturnCode.SEEK_NEXT_USING_HINT; -104 } -105} else { -106 hint = sortedPrefixes.first(); -107 return ReturnCode.SEEK_NEXT_USING_HINT; -108} -109 } -110 -111 public static Filter createFilterFromArguments(ArrayListbyte [] filterArguments) { -112byte [][] prefixes = new byte [filterArguments.size()][]; -113for (int i = 0 ; i filterArguments.size(); i++) { -114 byte [] columnPrefix = ParseFilter.removeQuotesFromByteArray(filterArguments.get(i)); -115 prefixes[i] = columnPrefix; -116} -117return new MultipleColumnPrefixFilter(prefixes); -118 } -119 -120 /** -121 * @return The filter serialized using pb -122 */ -123 @Override -124 public byte [] toByteArray() { -125 FilterProtos.MultipleColumnPrefixFilter.Builder builder = -126 FilterProtos.MultipleColumnPrefixFilter.newBuilder(); -127for (byte [] element : sortedPrefixes) { -128 if (element != null) builder.addSortedPrefixes(UnsafeByteOperations.unsafeWrap(element)); -129} -130return builder.build().toByteArray(); -131 } -132 -133 /** -134 * @param pbBytes A pb serialized {@link MultipleColumnPrefixFilter} instance -135 *
[23/51] [partial] hbase-site git commit: Published site at 7c1fad4992a169a35b4457e6f4afcb30d04406e9.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/74f60271/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTable.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTable.html b/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTable.html index ad86b01..57c6c7e 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTable.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTable.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -136,7 +136,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html private MetricsTableWrapperAggregate -tableWrapperAgg +wrapper @@ -178,6 +178,64 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html MetricsTableWrapperAggregate getTableWrapperAgg() + +void +incrSplitRequest(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) + + +void +incrSplitSuccess(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) + + +void +updateCompactionInputFileCount(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable, + booleanisMajor, + longc) + + +void +updateCompactionInputSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable, + booleanisMajor, + longbytes) + + +void +updateCompactionOutputFileCount(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable, + booleanisMajor, + longc) + + +void +updateCompactionOutputSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable, + booleanisMajor, + longbytes) + + +void +updateCompactionTime(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable, +booleanisMajor, +longt) + + +void +updateFlushMemstoreSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable, + longbytes) + + +void +updateFlushOutputSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable, + longbytes) + + +void +updateFlushTime(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable, + longt) + + +void +updateSplitTime(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable, + longt) + @@ -209,13 +267,13 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html private finalMetricsTableAggregateSource tableSourceAgg - + -tableWrapperAgg -privateMetricsTableWrapperAggregate tableWrapperAgg +wrapper +privateMetricsTableWrapperAggregate wrapper @@ -255,12 +313,125 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html - + getTableSourceAgg publicMetricsTableAggregateSourcegetTableSourceAgg() + + + + + +incrSplitRequest +publicvoidincrSplitRequest(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) + + + + + + + +incrSplitSuccess +publicvoidincrSplitSuccess(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) + + + + + + + +updateSplitTime +publicvoidupdateSplitTime(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable, +longt) + + + + + + + +updateFlushTime +publicvoidupdateFlushTime(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable, +longt) + + + + + + + +updateFlushMemstoreSize
[23/51] [partial] hbase-site git commit: Published site at 3afe9fb7e6ebfa71187cbe131558a83fae61cecd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html index c9128a4..c5b8dbd 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/RegionStates.html @@ -63,750 +63,734 @@ 055public class RegionStates { 056 private static final Logger LOG = LoggerFactory.getLogger(RegionStates.class); 057 -058 // TODO: need to be more specific, i.e, OPENING vs. OPEN, CLOSING vs. CLOSED. -059 static final State[] STATES_EXPECTED_ON_OPEN = new State[] { -060State.OPEN, // State may already be OPEN if we died after receiving the OPEN from regionserver -061// but before complete finish of AssignProcedure. HBASE-20100. -062State.OFFLINE, State.CLOSED, State.ABNORMALLY_CLOSED, // disable/offline -063State.SPLITTING, // ServerCrashProcedure -064State.OPENING, State.FAILED_OPEN, // already in-progress (retrying) -065State.MERGED, State.SPLITTING_NEW -066 }; -067 -068 static final State[] STATES_EXPECTED_ON_CLOSE = new State[] { -069State.SPLITTING, State.MERGING, State.OPENING, // ServerCrashProcedure -070State.OPEN, // enabled/open -071State.CLOSING // already in-progress (retrying) -072 }; -073 -074 // This comparator sorts the RegionStates by time stamp then Region name. -075 // Comparing by timestamp alone can lead us to discard different RegionStates that happen -076 // to share a timestamp. -077 private static class RegionStateStampComparator implements ComparatorRegionState { -078@Override -079public int compare(final RegionState l, final RegionState r) { -080 int stampCmp = Long.compare(l.getStamp(), r.getStamp()); -081 return stampCmp != 0 ? stampCmp : RegionInfo.COMPARATOR.compare(l.getRegion(), r.getRegion()); -082} -083 } -084 -085 public final static RegionStateStampComparator REGION_STATE_STAMP_COMPARATOR = -086 new RegionStateStampComparator(); -087 -088 // TODO: Replace the ConcurrentSkipListMaps -089 /** -090 * RegionName -- i.e. RegionInfo.getRegionName() -- as bytes to {@link RegionStateNode} -091 */ -092 private final ConcurrentSkipListMapbyte[], RegionStateNode regionsMap = -093 new ConcurrentSkipListMapbyte[], RegionStateNode(Bytes.BYTES_COMPARATOR); +058 // This comparator sorts the RegionStates by time stamp then Region name. +059 // Comparing by timestamp alone can lead us to discard different RegionStates that happen +060 // to share a timestamp. +061 private static class RegionStateStampComparator implements ComparatorRegionState { +062@Override +063public int compare(final RegionState l, final RegionState r) { +064 int stampCmp = Long.compare(l.getStamp(), r.getStamp()); +065 return stampCmp != 0 ? stampCmp : RegionInfo.COMPARATOR.compare(l.getRegion(), r.getRegion()); +066} +067 } +068 +069 public final static RegionStateStampComparator REGION_STATE_STAMP_COMPARATOR = +070 new RegionStateStampComparator(); +071 +072 // TODO: Replace the ConcurrentSkipListMaps +073 /** +074 * RegionName -- i.e. RegionInfo.getRegionName() -- as bytes to {@link RegionStateNode} +075 */ +076 private final ConcurrentSkipListMapbyte[], RegionStateNode regionsMap = +077 new ConcurrentSkipListMapbyte[], RegionStateNode(Bytes.BYTES_COMPARATOR); +078 +079 private final ConcurrentSkipListMapRegionInfo, RegionStateNode regionInTransition = +080new ConcurrentSkipListMapRegionInfo, RegionStateNode(RegionInfo.COMPARATOR); +081 +082 /** +083 * Regions marked as offline on a read of hbase:meta. Unused or at least, once +084 * offlined, regions have no means of coming on line again. TODO. +085 */ +086 private final ConcurrentSkipListMapRegionInfo, RegionStateNode regionOffline = +087new ConcurrentSkipListMapRegionInfo, RegionStateNode(); +088 +089 private final ConcurrentSkipListMapbyte[], RegionFailedOpen regionFailedOpen = +090new ConcurrentSkipListMapbyte[], RegionFailedOpen(Bytes.BYTES_COMPARATOR); +091 +092 private final ConcurrentHashMapServerName, ServerStateNode serverMap = +093 new ConcurrentHashMapServerName, ServerStateNode(); 094 -095 private final ConcurrentSkipListMapRegionInfo, RegionStateNode regionInTransition = -096new ConcurrentSkipListMapRegionInfo, RegionStateNode(RegionInfo.COMPARATOR); -097 -098 /** -099 * Regions marked as offline on a read of hbase:meta. Unused or at least, once -100 * offlined, regions have no means of coming on line again. TODO. -101 */ -102 private final
[23/51] [partial] hbase-site git commit: Published site at a452487a9b82bfd33bc10683c3f8b8ae74d58883.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cf79db0/checkstyle-aggregate.html -- diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html index 6098f6a..236f443 100644 --- a/checkstyle-aggregate.html +++ b/checkstyle-aggregate.html @@ -7,7 +7,7 @@ - + Apache HBase Checkstyle Results @@ -9838,12 +9838,12 @@ http://checkstyle.sourceforge.net/config_javadoc.html#JavadocTagContinuationIndentation;>JavadocTagContinuationIndentation offset: 2 -763 +764 Error http://checkstyle.sourceforge.net/config_javadoc.html#NonEmptyAtclauseDescription;>NonEmptyAtclauseDescription -3606 +3605 Error misc @@ -19479,7 +19479,7 @@ Error javadoc -NonEmptyAtclauseDescription +JavadocTagContinuationIndentation Javadoc comment at column 0 has parse error. Details: no viable alternative at input ' *' while parsing JAVADOC_TAG 117 @@ -23225,7 +23225,7 @@ imports ImportOrder Wrong order for 'org.apache.hadoop.hbase.exceptions.DeserializationException' import. -24 +25 org/apache/hadoop/hbase/client/ColumnFamilyDescriptor.java @@ -37613,55 +37613,55 @@ imports ImportOrder Wrong order for 'org.apache.hadoop.hbase.exceptions.DeserializationException' import. -27 +28 Error imports ImportOrder Wrong order for 'org.apache.hbase.thirdparty.com.google.common.base.Preconditions' import. -30 +31 Error blocks NeedBraces 'if' construct must use '{}'s. -56 +57 Error annotation MissingDeprecated Must include both @java.lang.Deprecated annotation and @deprecated Javadoc tag with description. -65 +66 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -103 +104 Error indentation Indentation 'throws' has incorrect indentation level 2, expected level should be 4. -107 +108 Error javadoc JavadocTagContinuationIndentation Line continuation have incorrect indentation level, expected level should be 2. -120 +121 Error blocks NeedBraces 'if' construct must use '{}'s. -124 +125 Error blocks NeedBraces 'if' construct must use '{}'s. -125 +126 org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java @@ -37676,79 +37676,79 @@ imports ImportOrder Wrong order for 'org.apache.hadoop.hbase.exceptions.DeserializationException' import. -28 +29 Error imports ImportOrder Wrong order for 'org.apache.hadoop.hbase.util.Bytes' import. -30 +31 Error sizes LineLength Line is longer than 100 characters (found 115). -38 +39 Error sizes LineLength Line is longer than 100 characters (found 110). -39 +40 Error blocks LeftCurly '{' at column 3 should be on the previous line. -60 +61 Error annotation MissingDeprecated Must include both @java.lang.Deprecated annotation and @deprecated Javadoc tag with description. -113 +114 Error blocks LeftCurly '{' at column 3 should be on the previous line. -121 +122 Error blocks LeftCurly '{' at column 3 should be on the previous line. -156 +157 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -188 +189 Error indentation Indentation 'throws' has incorrect indentation level 2, expected level should be 4. -192 +193 Error javadoc JavadocTagContinuationIndentation Line continuation have incorrect indentation level, expected level should be 2. -209 +210 Error blocks NeedBraces 'if' construct must use '{}'s. -213 +214 Error blocks NeedBraces 'if' construct must use '{}'s. -214 +215 org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java @@ -37763,73 +37763,73 @@ imports ImportOrder Wrong order for 'org.apache.hadoop.hbase.exceptions.DeserializationException' import. -29 +30 Error imports ImportOrder Wrong order for 'org.apache.hadoop.hbase.util.ByteBufferUtils' import. -31 +32 Error annotation MissingDeprecated Must include both @java.lang.Deprecated annotation and @deprecated Javadoc tag with description. -61 +62 Error blocks NeedBraces 'if' construct must use '{}'s. -120 +121 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -127 +128 Error indentation Indentation 'throws' has incorrect indentation level 2, expected level should be 4. -131 +132 Error javadoc JavadocTagContinuationIndentation Line continuation have incorrect indentation level, expected level should be 2. -144 +145 Error blocks NeedBraces 'if' construct must use '{}'s. -148 +149 Error indentation Indentation 'if' has incorrect indentation level 3, expected level should be 4. -148 +149 Error blocks NeedBraces 'if' construct must use '{}'s. -149 +150 Error indentation Indentation 'if' has incorrect indentation level 3, expected level should be 4. -149 +150 Error indentation Indentation 'method def' child has incorrect indentation level 3, expected level should be 4. -151 +152
[23/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html index db8431b..a8cb7c4 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html @@ -885,7766 +885,7797 @@ 877 * @return What the next sequence (edit) id should be. 878 * @throws IOException e 879 */ -880 private long initialize(final CancelableProgressable reporter) throws IOException { -881 -882//Refuse to open the region if there is no column family in the table -883if (htableDescriptor.getColumnFamilyCount() == 0) { -884 throw new DoNotRetryIOException("Table " + htableDescriptor.getTableName().getNameAsString()+ -885 " should have at least one column family."); -886} -887 -888MonitoredTask status = TaskMonitor.get().createStatus("Initializing region " + this); -889long nextSeqId = -1; -890try { -891 nextSeqId = initializeRegionInternals(reporter, status); -892 return nextSeqId; -893} finally { -894 // nextSeqid will be -1 if the initialization fails. -895 // At least it will be 0 otherwise. -896 if (nextSeqId == -1) { -897status.abort("Exception during region " + getRegionInfo().getRegionNameAsString() + -898 " initialization."); -899 } -900} -901 } -902 -903 private long initializeRegionInternals(final CancelableProgressable reporter, -904 final MonitoredTask status) throws IOException { -905if (coprocessorHost != null) { -906 status.setStatus("Running coprocessor pre-open hook"); -907 coprocessorHost.preOpen(); -908} -909 -910// Write HRI to a file in case we need to recover hbase:meta -911// Only the primary replica should write .regioninfo -912if (this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) { -913 status.setStatus("Writing region info on filesystem"); -914 fs.checkRegionInfoOnFilesystem(); -915} -916 -917// Initialize all the HStores -918status.setStatus("Initializing all the Stores"); -919long maxSeqId = initializeStores(reporter, status); -920this.mvcc.advanceTo(maxSeqId); -921if (ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) { -922 CollectionHStore stores = this.stores.values(); -923 try { -924// update the stores that we are replaying -925LOG.debug("replaying wal for " + this.getRegionInfo().getEncodedName()); -926 stores.forEach(HStore::startReplayingFromWAL); -927// Recover any edits if available. -928maxSeqId = Math.max(maxSeqId, -929 replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, status)); -930// Make sure mvcc is up to max. -931this.mvcc.advanceTo(maxSeqId); -932 } finally { -933LOG.debug("stopping wal replay for " + this.getRegionInfo().getEncodedName()); -934// update the stores that we are done replaying -935 stores.forEach(HStore::stopReplayingFromWAL); -936 } -937} -938this.lastReplayedOpenRegionSeqId = maxSeqId; -939 -940 this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this)); -941this.writestate.flushRequested = false; -942this.writestate.compacting.set(0); -943 -944if (this.writestate.writesEnabled) { -945 LOG.debug("Cleaning up temporary data for " + this.getRegionInfo().getEncodedName()); -946 // Remove temporary data left over from old regions -947 status.setStatus("Cleaning up temporary data from old regions"); -948 fs.cleanupTempDir(); -949} -950 -951if (this.writestate.writesEnabled) { -952 status.setStatus("Cleaning up detritus from prior splits"); -953 // Get rid of any splits or merges that were lost in-progress. Clean out -954 // these directories here on open. We may be opening a region that was -955 // being split but we crashed in the middle of it all. -956 LOG.debug("Cleaning up detritus for " + this.getRegionInfo().getEncodedName()); -957 fs.cleanupAnySplitDetritus(); -958 fs.cleanupMergesDir(); -959} +880 @VisibleForTesting +881 long initialize(final CancelableProgressable reporter) throws IOException { +882 +883//Refuse to open the region if there is no column family in the table +884if (htableDescriptor.getColumnFamilyCount() == 0) { +885 throw new DoNotRetryIOException("Table " + htableDescriptor.getTableName().getNameAsString()+ +886 " should have at
[23/51] [partial] hbase-site git commit: Published site at 63f2d3cbdc8151f5f61f33e0a078c51b9ac076a5.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ae6a80c/devapidocs/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.html index 617a7a2..db068f2 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/UnassignProcedure.html @@ -18,8 +18,8 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10}; -var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var methods = {"i0":42,"i1":42,"i2":42,"i3":42,"i4":42,"i5":42,"i6":42,"i7":42,"i8":42,"i9":42,"i10":42,"i11":42}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"],32:["t6","Deprecated Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; var tableTab = "tableTab"; @@ -49,8 +49,8 @@ var activeTableTab = "activeTableTab"; -PrevClass -NextClass +PrevClass +NextClass Frames @@ -122,32 +122,19 @@ var activeTableTab = "activeTableTab"; https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableProcedureMasterProcedureEnv, TableProcedureInterface, RemoteProcedureDispatcher.RemoteProcedureMasterProcedureEnv,ServerName +Deprecated. +Do not use any more. + -@InterfaceAudience.Private -public class UnassignProcedure +https://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true; title="class or interface in java.lang">@Deprecated + @InterfaceAudience.Private +public class UnassignProcedure extends RegionTransitionProcedure -Procedure that describes the unassignment of a single region. - There can only be one RegionTransitionProcedure -- i.e. an assign or an unassign -- per region - running at a time, since each procedure takes a lock on the region. - - The Unassign starts by placing a "close region" request in the Remote Dispatcher - queue, and the procedure will then go into a "waiting state" (suspend). - The Remote Dispatcher will batch the various requests for that server and - they will be sent to the RS for execution. - The RS will complete the open operation by calling master.reportRegionStateTransition(). - The AM will intercept the transition report, and notify this procedure. - The procedure will wakeup and finish the unassign by publishing its new state on meta. - If we are unable to contact the remote regionserver whether because of ConnectException - or socket timeout, we will call expire on the server we were trying to contact. We will remain - in suspended state waiting for a wake up from the ServerCrashProcedure that is processing the - failed server. The basic idea is that if we notice a crashed server, then we have a - responsibility; i.e. we should not let go of the region until we are sure the server that was - hosting has had its crash processed. If we let go of the region before then, an assign might - run before the logs have been split which would make for data loss. - - TODO: Rather than this tricky coordination between SCP and this Procedure, instead, work on - returning a SCP as our subprocedure; probably needs work on the framework to do this, - especially if the SCP already created. +Leave here only for checking if we can successfully start the master. + +See Also: +TransitRegionStateProcedure + @@ -191,28 +178,26 @@ extends protected ServerName destinationServer -The Server we will subsequently assign the region too (can be null). - +Deprecated. + private boolean -force +force +Deprecated. + protected ServerName hostingServer -Where to send the unassign RPC. - +Deprecated. + -private static org.slf4j.Logger -LOG - - private boolean removeAfterUnassigning -Whether deleting the region from in-memory states after unassigning the region. - +Deprecated. + @@ -243,26 +228,9 @@ extends Constructor and Description -UnassignProcedure() - - -UnassignProcedure(RegionInforegionInfo, - ServerNamehostingServer, - booleanforce, - booleanremoveAfterUnassigning) - - -UnassignProcedure(RegionInforegionInfo, - ServerNamehostingServer, - ServerNamedestinationServer, - booleanforce) - - -UnassignProcedure(RegionInforegionInfo, - ServerNamehostingServer, - ServerNamedestinationServer, - booleanforce, - booleanremoveAfterUnassigning) +UnassignProcedure() +Deprecated. + @@ -274,7 +242,7 @@ extends -All MethodsInstance MethodsConcrete Methods +All
[23/51] [partial] hbase-site git commit: Published site at 092efb42749bf7fc6ad338c96aae8e7b9d3a2c74.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f3d62514/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/TsvImporterTextMapper.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/TsvImporterTextMapper.html b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/TsvImporterTextMapper.html index 8359449..24080ca 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/TsvImporterTextMapper.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/TsvImporterTextMapper.html @@ -36,104 +36,106 @@ 028import org.apache.hadoop.mapreduce.Counter; 029import org.apache.yetus.audience.InterfaceAudience; 030import org.apache.hadoop.conf.Configuration; -031 -032/** -033 * Write table content out to map output files. -034 */ -035@InterfaceAudience.Public -036public class TsvImporterTextMapper -037extends MapperLongWritable, Text, ImmutableBytesWritable, Text -038{ -039 -040 /** Column seperator */ -041 private String separator; -042 -043 /** Should skip bad lines */ -044 private boolean skipBadLines; -045 private Counter badLineCount; -046 private boolean logBadLines; -047 -048 private ImportTsv.TsvParser parser; +031import org.slf4j.Logger; +032import org.slf4j.LoggerFactory; +033 +034/** +035 * Write table content out to map output files. +036 */ +037@InterfaceAudience.Public +038public class TsvImporterTextMapper +039extends MapperLongWritable, Text, ImmutableBytesWritable, Text { +040 private static final Logger LOG = LoggerFactory.getLogger(TsvImporterTextMapper.class); +041 +042 /** Column seperator */ +043 private String separator; +044 +045 /** Should skip bad lines */ +046 private boolean skipBadLines; +047 private Counter badLineCount; +048 private boolean logBadLines; 049 -050 public boolean getSkipBadLines() { -051return skipBadLines; -052 } -053 -054 public Counter getBadLineCount() { -055return badLineCount; -056 } -057 -058 public void incrementBadLineCount(int count) { -059this.badLineCount.increment(count); -060 } -061 -062 /** -063 * Handles initializing this class with objects specific to it (i.e., the parser). -064 * Common initialization that might be leveraged by a subclass is done in -065 * codedoSetup/code. Hence a subclass may choose to override this method -066 * and call codedoSetup/code as well before handling it's own custom params. -067 * -068 * @param context -069 */ -070 @Override -071 protected void setup(Context context) { -072doSetup(context); -073 -074Configuration conf = context.getConfiguration(); +050 private ImportTsv.TsvParser parser; +051 +052 public boolean getSkipBadLines() { +053return skipBadLines; +054 } +055 +056 public Counter getBadLineCount() { +057return badLineCount; +058 } +059 +060 public void incrementBadLineCount(int count) { +061this.badLineCount.increment(count); +062 } +063 +064 /** +065 * Handles initializing this class with objects specific to it (i.e., the parser). +066 * Common initialization that might be leveraged by a subclass is done in +067 * codedoSetup/code. Hence a subclass may choose to override this method +068 * and call codedoSetup/code as well before handling it's own custom params. +069 * +070 * @param context +071 */ +072 @Override +073 protected void setup(Context context) { +074doSetup(context); 075 -076parser = new ImportTsv.TsvParser(conf.get(ImportTsv.COLUMNS_CONF_KEY), separator); -077if (parser.getRowKeyColumnIndex() == -1) { -078 throw new RuntimeException("No row key column specified"); -079} -080 } -081 -082 /** -083 * Handles common parameter initialization that a subclass might want to leverage. -084 * @param context -085 */ -086 protected void doSetup(Context context) { -087Configuration conf = context.getConfiguration(); -088 -089// If a custom separator has been used, -090// decode it back from Base64 encoding. -091separator = conf.get(ImportTsv.SEPARATOR_CONF_KEY); -092if (separator == null) { -093 separator = ImportTsv.DEFAULT_SEPARATOR; -094} else { -095 separator = new String(Base64.getDecoder().decode(separator)); -096} -097 -098skipBadLines = context.getConfiguration().getBoolean(ImportTsv.SKIP_LINES_CONF_KEY, true); -099logBadLines = context.getConfiguration().getBoolean(ImportTsv.LOG_BAD_LINES_CONF_KEY, false); -100badLineCount = context.getCounter("ImportTsv", "Bad Lines"); -101 } -102 -103 /** -104 * Convert a line of TSV text into an HBase table row. -105 */ -106 @Override -107 public void map(LongWritable offset, Text value, Context context) throws IOException { -108try { -109 PairInteger,Integer rowKeyOffests = parser.parseRowKey(value.getBytes(), value.getLength()); -110 ImmutableBytesWritable rowKey = new ImmutableBytesWritable( -111
[23/51] [partial] hbase-site git commit: Published site at 613d831429960348dc42c3bdb6ea5d31be15c81c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html index b7b4236..3d1edb3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html @@ -259,1863 +259,1867 @@ 251 * + Metadata! + = See note on BLOCK_METADATA_SPACE above. 252 * ++ 253 * /code -254 * @see #serialize(ByteBuffer) +254 * @see #serialize(ByteBuffer, boolean) 255 */ -256 static final CacheableDeserializerCacheable BLOCK_DESERIALIZER = -257 new CacheableDeserializerCacheable() { -258@Override -259public HFileBlock deserialize(ByteBuff buf, boolean reuse, MemoryType memType) -260throws IOException { -261 // The buf has the file block followed by block metadata. -262 // Set limit to just before the BLOCK_METADATA_SPACE then rewind. -263 buf.limit(buf.limit() - BLOCK_METADATA_SPACE).rewind(); -264 // Get a new buffer to pass the HFileBlock for it to 'own'. -265 ByteBuff newByteBuff; -266 if (reuse) { -267newByteBuff = buf.slice(); -268 } else { -269int len = buf.limit(); -270newByteBuff = new SingleByteBuff(ByteBuffer.allocate(len)); -271newByteBuff.put(0, buf, buf.position(), len); -272 } -273 // Read out the BLOCK_METADATA_SPACE content and shove into our HFileBlock. -274 buf.position(buf.limit()); -275 buf.limit(buf.limit() + HFileBlock.BLOCK_METADATA_SPACE); -276 boolean usesChecksum = buf.get() == (byte) 1; -277 long offset = buf.getLong(); -278 int nextBlockOnDiskSize = buf.getInt(); -279 HFileBlock hFileBlock = -280 new HFileBlock(newByteBuff, usesChecksum, memType, offset, nextBlockOnDiskSize, null); -281 return hFileBlock; -282} -283 -284@Override -285public int getDeserialiserIdentifier() { -286 return DESERIALIZER_IDENTIFIER; -287} -288 -289@Override -290public HFileBlock deserialize(ByteBuff b) throws IOException { -291 // Used only in tests -292 return deserialize(b, false, MemoryType.EXCLUSIVE); -293} -294 }; -295 -296 private static final int DESERIALIZER_IDENTIFIER; -297 static { -298DESERIALIZER_IDENTIFIER = -299 CacheableDeserializerIdManager.registerDeserializer(BLOCK_DESERIALIZER); -300 } -301 -302 /** -303 * Copy constructor. Creates a shallow copy of {@code that}'s buffer. -304 */ -305 private HFileBlock(HFileBlock that) { -306this(that, false); -307 } -308 -309 /** -310 * Copy constructor. Creates a shallow/deep copy of {@code that}'s buffer as per the boolean -311 * param. -312 */ -313 private HFileBlock(HFileBlock that, boolean bufCopy) { -314init(that.blockType, that.onDiskSizeWithoutHeader, -315 that.uncompressedSizeWithoutHeader, that.prevBlockOffset, -316that.offset, that.onDiskDataSizeWithHeader, that.nextBlockOnDiskSize, that.fileContext); -317if (bufCopy) { -318 this.buf = new SingleByteBuff(ByteBuffer.wrap(that.buf.toBytes(0, that.buf.limit(; -319} else { -320 this.buf = that.buf.duplicate(); -321} -322 } -323 -324 /** -325 * Creates a new {@link HFile} block from the given fields. This constructor -326 * is used only while writing blocks and caching, -327 * and is sitting in a byte buffer and we want to stuff the block into cache. -328 * -329 * pTODO: The caller presumes no checksumming -330 * required of this block instance since going into cache; checksum already verified on -331 * underlying block data pulled in from filesystem. Is that correct? What if cache is SSD? +256 public static final CacheableDeserializerCacheable BLOCK_DESERIALIZER = new BlockDeserializer(); +257 +258 public static final class BlockDeserializer implements CacheableDeserializerCacheable { +259private BlockDeserializer() { +260} +261 +262@Override +263public HFileBlock deserialize(ByteBuff buf, boolean reuse, MemoryType memType) +264throws IOException { +265 // The buf has the file block followed by block metadata. +266 // Set limit to just before the BLOCK_METADATA_SPACE then rewind. +267 buf.limit(buf.limit() - BLOCK_METADATA_SPACE).rewind(); +268 // Get a new buffer to pass the HFileBlock for it to 'own'. +269 ByteBuff newByteBuff; +270 if (reuse) { +271newByteBuff = buf.slice(); +272 } else { +273int len = buf.limit(); +274newByteBuff = new SingleByteBuff(ByteBuffer.allocate(len)); +275newByteBuff.put(0, buf, buf.position(), len);
[23/51] [partial] hbase-site git commit: Published site at ba5d1c1f28301adc99019d9d6c4a04fac98ae511.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/804782f0/devapidocs/org/apache/hadoop/hbase/master/replication/RefreshPeerProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/replication/RefreshPeerProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/replication/RefreshPeerProcedure.html index 4bde540..0a56e97 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/replication/RefreshPeerProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/replication/RefreshPeerProcedure.html @@ -334,7 +334,7 @@ implements Procedure -acquireLock, addStackIndex, afterReplay, beforeReplay, compareTo, completionCleanup, doAcquireLock, d oExecute, doReleaseLock, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isFailed, isFinished, isInitializing, i sRunnable, isSuccess, isWaiting, isYieldAfterExecutionStep, releaseLock, removeStackIndex, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, shouldWaitClientAck, toString, toStringClass, toStringClassDetails, toStringDetails, toStringSimpleSB, toStringState, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, wasExecuted +acquireLock, addStackIndex, afterReplay, beforeReplay, compareTo, completionCleanup, doExecute, doRollba ck, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedT ime, getTimeout, getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, isYieldAfterExecutionStep, releaseLock, removeStackIndex, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, shouldWaitClientAck, toString, toStringClass, toStringClassDetails, toStringDetails, toStringSimpleSB, toStringState, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, waitInitialized, wasExecuted http://git-wip-us.apache.org/repos/asf/hbase-site/blob/804782f0/devapidocs/org/apache/hadoop/hbase/master/replication/RemovePeerProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/replication/RemovePeerProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/replication/RemovePeerProcedure.html index b7f19fe..742e018 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/replication/RemovePeerProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/replication/RemovePeerProcedure.html @@ -299,7 +299,7 @@ extends AbstractPeerProcedure -acquireLock, getLatch, getPeerId, hasLock, holdLock, refreshPeer, releaseLock, rollbackState +acquireLock, getLatch, getPeerId, holdLock, refreshPeer, releaseLock, rollbackState, waitInitialized @@ -313,7 +313,7 @@ extends Procedure -addStackIndex, afterReplay, beforeReplay, compareTo, completionCleanup, doAcquireLock, doExecute, d oReleaseLock, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutTimestamp, hasChildren, hasException, hasOwner, hasParent, hasTimeout, haveSameParent, incChildrenLatch, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, removeStackIndex, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, < a href="../../../../../../org/apache/hadoop/hbase/procedure2/Procedure.html#setTimeout-int-">setTimeout, setTimeoutFailure, shouldWaitClientAck, toString, toStringClass, toStringClassDetails, toStringDetails, toStringSimpleSB, updateMetricsOnFinish,
[23/51] [partial] hbase-site git commit: Published site at b4759ce6e72f50ccd9d410bd5917dc5a515414f1.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/df8fd1d3/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.LocalityBasedCandidateGenerator.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.LocalityBasedCandidateGenerator.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.LocalityBasedCandidateGenerator.html index 233dba3..91b9055 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.LocalityBasedCandidateGenerator.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.LocalityBasedCandidateGenerator.html @@ -540,1205 +540,1204 @@ 532 sm.getRegionMetrics().forEach((byte[] regionName, RegionMetrics rm) - { 533DequeBalancerRegionLoad rLoads = oldLoads.get(Bytes.toString(regionName)); 534if (rLoads == null) { -535 // There was nothing there -536 rLoads = new ArrayDeque(); -537} else if (rLoads.size() = numRegionLoadsToRemember) { -538 rLoads.remove(); -539} -540rLoads.add(new BalancerRegionLoad(rm)); -541 loads.put(Bytes.toString(regionName), rLoads); -542 }); -543}); -544 -545for(CostFromRegionLoadFunction cost : regionLoadFunctions) { -546 cost.setLoads(loads); -547} -548 } -549 -550 protected void initCosts(Cluster cluster) { -551for (CostFunction c:costFunctions) { -552 c.init(cluster); -553} -554 } -555 -556 protected void updateCostsWithAction(Cluster cluster, Action action) { -557for (CostFunction c : costFunctions) { -558 c.postAction(action); -559} -560 } -561 -562 /** -563 * Get the names of the cost functions -564 */ -565 public String[] getCostFunctionNames() { -566if (costFunctions == null) return null; -567String[] ret = new String[costFunctions.length]; -568for (int i = 0; i costFunctions.length; i++) { -569 CostFunction c = costFunctions[i]; -570 ret[i] = c.getClass().getSimpleName(); -571} -572 -573return ret; -574 } -575 -576 /** -577 * This is the main cost function. It will compute a cost associated with a proposed cluster -578 * state. All different costs will be combined with their multipliers to produce a double cost. -579 * -580 * @param cluster The state of the cluster -581 * @param previousCost the previous cost. This is used as an early out. -582 * @return a double of a cost associated with the proposed cluster state. This cost is an -583 * aggregate of all individual cost functions. -584 */ -585 protected double computeCost(Cluster cluster, double previousCost) { -586double total = 0; -587 -588for (int i = 0; i costFunctions.length; i++) { -589 CostFunction c = costFunctions[i]; -590 this.tempFunctionCosts[i] = 0.0; -591 -592 if (c.getMultiplier() = 0) { -593continue; -594 } -595 -596 Float multiplier = c.getMultiplier(); -597 Double cost = c.cost(); -598 -599 this.tempFunctionCosts[i] = multiplier*cost; -600 total += this.tempFunctionCosts[i]; -601 -602 if (total previousCost) { -603break; -604 } -605} -606 -607return total; -608 } -609 -610 /** Generates a candidate action to be applied to the cluster for cost function search */ -611 abstract static class CandidateGenerator { -612abstract Cluster.Action generate(Cluster cluster); -613 -614/** -615 * From a list of regions pick a random one. Null can be returned which -616 * {@link StochasticLoadBalancer#balanceCluster(Map)} recognize as signal to try a region move -617 * rather than swap. -618 * -619 * @param clusterThe state of the cluster -620 * @param server index of the server -621 * @param chanceOfNoSwap Chance that this will decide to try a move rather -622 * than a swap. -623 * @return a random {@link RegionInfo} or null if an asymmetrical move is -624 * suggested. -625 */ -626protected int pickRandomRegion(Cluster cluster, int server, double chanceOfNoSwap) { -627 // Check to see if this is just a move. -628 if (cluster.regionsPerServer[server].length == 0 || RANDOM.nextFloat() chanceOfNoSwap) { -629// signal a move only. -630return -1; -631 } -632 int rand = RANDOM.nextInt(cluster.regionsPerServer[server].length); -633 return cluster.regionsPerServer[server][rand]; -634 -635} -636protected int pickRandomServer(Cluster cluster) { -637 if (cluster.numServers 1) { -638return -1; -639 } -640 -641 return RANDOM.nextInt(cluster.numServers); -642} -643 -644protected int pickRandomRack(Cluster cluster) { -645 if (cluster.numRacks
[23/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0c6f447e/apidocs/org/apache/hadoop/hbase/NotServingRegionException.html -- diff --git a/apidocs/org/apache/hadoop/hbase/NotServingRegionException.html b/apidocs/org/apache/hadoop/hbase/NotServingRegionException.html index 1665f28..fd82164 100644 --- a/apidocs/org/apache/hadoop/hbase/NotServingRegionException.html +++ b/apidocs/org/apache/hadoop/hbase/NotServingRegionException.html @@ -1,6 +1,6 @@ http://www.w3.org/TR/html4/loose.dtd;> - + @@ -20,38 +20,38 @@ //--> -æ¨çæµè§å¨å·²ç¦ç¨ JavaScriptã +JavaScript is disabled on your browser. -è·³è¿å¯¼èªé¾æ¥ +Skip navigation links - -æ¦è§ -ç¨åºå -ç±» -ä½¿ç¨ -æ -å·²è¿æ¶ -ç´¢å¼ -å¸®å© + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help -ä¸ä¸ä¸ªç±» -ä¸ä¸ä¸ªç±» +PrevClass +NextClass -æ¡æ¶ -æ æ¡æ¶ +Frames +NoFrames -ææç±» +AllClassesorg.apache.hadoop.hbase-ç±» NotServingRegionException
+Class NotServingRegionException
-
- https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true" title="java.langä¸çç±»ææ¥å£">java.lang.Object
+- https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true" title="class or interface in java.lang">java.lang.Object
-
- https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true" title="java.langä¸çç±»ææ¥å£">java.lang.Throwable
+- https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true" title="class or interface in java.lang">java.lang.Throwable
-
- https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true" title="java.langä¸çç±»ææ¥å£">java.lang.Exception
+- https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true" title="class or interface in java.lang">java.lang.Exception
-
- https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="java.ioä¸çç±»ææ¥å£">java.io.IOException
+- https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">java.io.IOException
- org.apache.hadoop.hbase.NotServingRegionException
@@ -117,19 +117,19 @@
-
- ææå·²å®ç°çæ¥å£:
-
[23/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5427a45e/apidocs/org/apache/hadoop/hbase/NamespaceExistException.html -- diff --git a/apidocs/org/apache/hadoop/hbase/NamespaceExistException.html b/apidocs/org/apache/hadoop/hbase/NamespaceExistException.html index e655215..ba2dc6f 100644 --- a/apidocs/org/apache/hadoop/hbase/NamespaceExistException.html +++ b/apidocs/org/apache/hadoop/hbase/NamespaceExistException.html @@ -1,6 +1,6 @@ http://www.w3.org/TR/html4/loose.dtd;> - + @@ -20,38 +20,38 @@ //--> -JavaScript is disabled on your browser. +æ¨çæµè§å¨å·²ç¦ç¨ JavaScriptã -Skip navigation links +è·³è¿å¯¼èªé¾æ¥ - -Overview -Package -Class -Use -Tree -Deprecated -Index -Help + +æ¦è§ +ç¨åºå +ç±» +ä½¿ç¨ +æ +å·²è¿æ¶ +ç´¢å¼ +å¸®å© -PrevClass -NextClass +ä¸ä¸ä¸ªç±» +ä¸ä¸ä¸ªç±» -Frames -NoFrames +æ¡æ¶ +æ æ¡æ¶ -AllClasses +ææç±»org.apache.hadoop.hbase-Class NamespaceExistException
+ç±» NamespaceExistException
-
- https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true" title="class or interface in java.lang">java.lang.Object
+- https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true" title="java.langä¸çç±»ææ¥å£">java.lang.Object
-
- https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true" title="class or interface in java.lang">java.lang.Throwable
+- https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true" title="java.langä¸çç±»ææ¥å£">java.lang.Throwable
-
- https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true" title="class or interface in java.lang">java.lang.Exception
+- https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true" title="java.langä¸çç±»ææ¥å£">java.lang.Exception
-
- https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">java.io.IOException
+- https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="java.ioä¸çç±»ææ¥å£">java.io.IOException
-
- org.apache.hadoop.hbase.HBaseIOException
+
[23/51] [partial] hbase-site git commit: Published site at 0f23784182ab88649de340d75804e0ff20dcd0fc.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bcb555af/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.html b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.html index 432cba4..8851d0f 100644 --- a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.html +++ b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public class ReplicationSource +public class ReplicationSource extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements ReplicationSourceInterface Class that handles the source of a replication stream. @@ -369,55 +369,65 @@ implements +private long +getFileSize(org.apache.hadoop.fs.PathcurrentPath) + + ReplicationPeer getPeer() Get the replication peer instance. - + https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String getQueueId() Get the queue id that the source is replicating to - + (package private) ReplicationQueueStorage getQueueStorage() - + ReplicationEndpoint getReplicationEndpoint() - + (package private) Server getServer() - + ServerName getServerWALsBelongTo() The queue of WALs only belong to one region server. - + ReplicationSourceManager getSourceManager() - + MetricsSource getSourceMetrics() - + https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String getStats() Get a string representation of the current statistics for this source - + WALFileLengthProvider getWALFileLengthProvider() - + +https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,ReplicationStatus +getWalGroupStatus() +get the stat of replication for each wal group. + + + void init(org.apache.hadoop.conf.Configurationconf, org.apache.hadoop.fs.FileSystemfs, @@ -432,77 +442,77 @@ implements Instantiation method used by region servers - + private void initAndStartReplicationEndpoint(ReplicationEndpointreplicationEndpoint) - + private void initialize() - + private void initializeWALEntryFilter(https://docs.oracle.com/javase/8/docs/api/java/util/UUID.html?is-external=true; title="class or interface in java.util">UUIDpeerClusterId) - + boolean isSourceActive() - + void postShipEdits(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListWAL.Entryentries, intbatchSize) Call this after the shipper thread ship some entries to peer cluster. - + (package private) void removeWorker(ReplicationSourceShipperworker) - + protected boolean sleepForRetries(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringmsg, intsleepMultiplier) Do the sleeping logic - + void startup() Start the replication - + void terminate(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringreason) End the replication - + void terminate(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringreason, https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">Exceptioncause) End the replication - + void terminate(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
[23/51] [partial] hbase-site git commit: Published site at 85b41f36e01214b6485c9352875c84ebf877dab3.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a5c66de0/devapidocs/org/apache/hadoop/hbase/security/access/class-use/Permission.Action.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/security/access/class-use/Permission.Action.html b/devapidocs/org/apache/hadoop/hbase/security/access/class-use/Permission.Action.html index a1f615f..9ca96d9 100644 --- a/devapidocs/org/apache/hadoop/hbase/security/access/class-use/Permission.Action.html +++ b/devapidocs/org/apache/hadoop/hbase/security/access/class-use/Permission.Action.html @@ -506,10 +506,47 @@ the order they are declared. +static boolean +AccessControlClient.hasPermission(Connectionconnection, + https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringtableName, + byte[]columnFamily, + byte[]columnQualifier, + https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringuserName, + Permission.Action...actions) +Validates whether specified user has permission to perform actions on the mentioned table, + column family or column qualifier. + + + +static boolean +AccessControlClient.hasPermission(Connectionconnection, + https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringtableName, + https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringcolumnFamily, + https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringcolumnQualifier, + https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringuserName, + Permission.Action...actions) +Validates whether specified user has permission to perform actions on the mentioned table, + column family or column qualifier. + + + +static boolean +AccessControlUtil.hasPermission(com.google.protobuf.RpcControllercontroller, + org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterfaceprotocol, + TableNametableName, + byte[]columnFamily, + byte[]columnQualifier, + https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringuserName, + Permission.Action[]actions) +Validates whether specified user has permission to perform actions on the mentioned table, + column family or column qualifier. + + + boolean Permission.implies(Permission.Actionaction) - + boolean TablePermission.implies(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringnamespace, Permission.Actionaction) @@ -517,7 +554,7 @@ the order they are declared. instance. - + boolean TablePermission.implies(TableNametable, byte[]family, @@ -527,7 +564,7 @@ the order they are declared. instance. - + boolean TablePermission.implies(TableNametable, KeyValuekv, @@ -536,7 +573,7 @@ the order they are declared. the given table and key value. - + boolean TablePermission.matchesFamily(TableNametable, byte[]family, @@ -545,7 +582,7 @@ the order they are declared. family at least. - + boolean TablePermission.matchesFamilyQualifier(TableNametable, byte[]family, @@ -554,7 +591,7 @@ the order they are declared. Returns if the given permission matches the given qualifier. - + boolean TableAuthManager.matchPermission(Useruser, TableNametable, @@ -562,7 +599,7 @@ the order they are declared. byte[]qualifier, Permission.Actionaction) - + boolean TableAuthManager.matchPermission(Useruser, TableNametable, @@ -572,7 +609,7 @@ the order they are declared. to the column family portion of a permission. - + private AuthResult AccessController.permissionGranted(AccessController.OpTypeopType, Useruser, @@ -583,25 +620,26 @@ the order they are declared. against the given set of row data. - + private AuthResult -AccessController.permissionGranted(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringrequest, +AccessController.permissionGranted(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringrequest, Useruser, Permission.ActionpermRequest, RegionCoprocessorEnvironmente, +
[23/51] [partial] hbase-site git commit: Published site at 6198e1fc7dfa85c3bc6b2855f9a5fb5f4b2354ff.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb5d2c62/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html index 37f4bd5..c84d49c 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html +++ b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i 109":10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":10,"i145":10,"i146":10,"i147":10,"i148":10,"i149":10,"i150":10,"i151":10,"i152":10,"i153":10,"i154":10,"i155":10,"i156":10,"i157":10,"i158":10,"i159":10,"i160":10,"i161":10,"i162":10,"i163":10,"i164":10,"i165":10,"i166":10,"i167":10,"i168":10,"i169":10,"i170":10,"i171":10,"i172":10,"i173":10,"i174":10,"i175":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i 109":10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":10,"i145":10,"i146":10,"i147":10,"i148":10,"i149":10,"i150":10,"i151":10,"i152":10,"i153":10,"i154":10,"i155":10,"i156":10,"i157":10,"i158":10,"i159":10,"i160":10,"i161":10,"i162":10,"i163":10,"i164":10,"i165":10,"i166":10,"i167":10,"i168":10,"i169":10,"i170":10,"i171":10,"i172":10,"i173":10,"i174":10,"i175":10,"i176":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -class RawAsyncHBaseAdmin +class RawAsyncHBaseAdmin extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements AsyncAdmin The implementation of AsyncAdmin. @@ -1355,49 +1355,56 @@ implements https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void +transitReplicationPeerSyncReplicationState(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringpeerId, +
[23/51] [partial] hbase-site git commit: Published site at 14087cc919da9f2e0b1a68f701f6365ad9d1d71f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/55ce8d97/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html index ccbd322..d044d2e 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":9,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10}; +var methods = {"i0":9,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -191,10 +191,6 @@ extends modifiedTableDescriptor -private https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean -traceEnabled - - private TableDescriptor unmodifiedTableDescriptor @@ -288,103 +284,85 @@ extends private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -getOpenRegionInfoList(MasterProcedureEnvenv) -Fetches all open or soon to be open Regions for a table. - - - -private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo getRegionInfoList(MasterProcedureEnvenv) Fetches all Regions for a table. - + protected org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableState getState(intstateId) Convert an ordinal (or state id) to an Enum (or more descriptive) state object. - + protected int getStateId(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStatestate) Convert the Enum (or more descriptive) state object to an ordinal (or state id). - + TableName getTableName() - + TableProcedureInterface.TableOperationType getTableOperationType() Given an operation type we can take decisions about what to do with pending operations. - + private void initilize() - + protected boolean isRollbackSupported(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStatestate) Used by the default implementation of abort() to know if the current state can be aborted and rollback can be triggered. - -private https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean -isTraceEnabled() -The procedure could be restarted from a different machine. - - - + private void postModify(MasterProcedureEnvenv, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStatestate) Action after modifying table. - + private void preModify(MasterProcedureEnvenv, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStatestate) Action before modifying table. - + private void prepareModify(MasterProcedureEnvenv) Check conditions before any real action of modifying a table. - -private void -restoreTableDescriptor(MasterProcedureEnvenv) -Undo the descriptor change (for rollback) - - - + protected void rollbackState(MasterProcedureEnvenv, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStatestate) called to perform the rollback of the specified state - + private void runCoprocessorAction(MasterProcedureEnvenv, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStatestate) Coprocessor Action. - + protected void serializeStateData(ProcedureStateSerializerserializer) The user-level code of the procedure may have some state to persist (e.g. - + private void updateReplicaColumnsIfNeeded(MasterProcedureEnvenv, TableDescriptoroldTableDescriptor, @@ -392,7 +370,7 @@ extends update replica column families if necessary. - + private void updateTableDescriptor(MasterProcedureEnvenv) Update descriptor @@ -471,21 +449,12 @@ extends - + deleteColumnFamilyInModify privateboolean deleteColumnFamilyInModify - - - - - -traceEnabled -privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean traceEnabled - - @@ -500,7
[23/51] [partial] hbase-site git commit: Published site at 72784c2d836a4b977667449d3adec5e8d15453f5.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2b11656f/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html index b6e7636..592c2cc 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html @@ -356,3901 +356,3924 @@ 348 public FutureVoid modifyTableAsync(TableDescriptor td) throws IOException { 349ModifyTableResponse response = executeCallable( 350 new MasterCallableModifyTableResponse(getConnection(), getRpcControllerFactory()) { -351@Override -352protected ModifyTableResponse rpcCall() throws Exception { -353 setPriority(td.getTableName()); -354 ModifyTableRequest request = RequestConverter.buildModifyTableRequest( -355td.getTableName(), td, ng.getNonceGroup(), ng.newNonce()); -356 return master.modifyTable(getRpcController(), request); -357} -358 }); -359return new ModifyTableFuture(this, td.getTableName(), response); -360 } -361 -362 @Override -363 public ListTableDescriptor listTableDescriptorsByNamespace(byte[] name) throws IOException { -364return executeCallable(new MasterCallableListTableDescriptor(getConnection(), -365getRpcControllerFactory()) { -366 @Override -367 protected ListTableDescriptor rpcCall() throws Exception { -368return master.listTableDescriptorsByNamespace(getRpcController(), -369 ListTableDescriptorsByNamespaceRequest.newBuilder() -370 .setNamespaceName(Bytes.toString(name)).build()) -371.getTableSchemaList() -372.stream() -373 .map(ProtobufUtil::toTableDescriptor) -374 .collect(Collectors.toList()); -375 } -376}); -377 } -378 -379 @Override -380 public ListTableDescriptor listTableDescriptors(ListTableName tableNames) throws IOException { -381return executeCallable(new MasterCallableListTableDescriptor(getConnection(), -382getRpcControllerFactory()) { -383 @Override -384 protected ListTableDescriptor rpcCall() throws Exception { -385GetTableDescriptorsRequest req = -386 RequestConverter.buildGetTableDescriptorsRequest(tableNames); -387 return ProtobufUtil.toTableDescriptorList(master.getTableDescriptors(getRpcController(), -388 req)); -389 } -390}); -391 } -392 -393 @Override -394 public ListRegionInfo getRegions(final ServerName sn) throws IOException { -395AdminService.BlockingInterface admin = this.connection.getAdmin(sn); -396// TODO: There is no timeout on this controller. Set one! -397HBaseRpcController controller = rpcControllerFactory.newController(); -398return ProtobufUtil.getOnlineRegions(controller, admin); -399 } -400 -401 @Override -402 public ListRegionInfo getRegions(TableName tableName) throws IOException { -403if (TableName.isMetaTableName(tableName)) { -404 return Arrays.asList(RegionInfoBuilder.FIRST_META_REGIONINFO); -405} else { -406 return MetaTableAccessor.getTableRegions(connection, tableName, true); -407} -408 } -409 -410 private static class AbortProcedureFuture extends ProcedureFutureBoolean { -411private boolean isAbortInProgress; -412 -413public AbortProcedureFuture( -414final HBaseAdmin admin, -415final Long procId, -416final Boolean abortProcResponse) { -417 super(admin, procId); -418 this.isAbortInProgress = abortProcResponse; -419} -420 -421@Override -422public Boolean get(long timeout, TimeUnit unit) -423throws InterruptedException, ExecutionException, TimeoutException { -424 if (!this.isAbortInProgress) { -425return false; -426 } -427 super.get(timeout, unit); -428 return true; -429} -430 } -431 -432 /** @return Connection used by this object. */ -433 @Override -434 public Connection getConnection() { -435return connection; -436 } -437 -438 @Override -439 public boolean tableExists(final TableName tableName) throws IOException { -440return executeCallable(new RpcRetryingCallableBoolean() { -441 @Override -442 protected Boolean rpcCall(int callTimeout) throws Exception { -443return MetaTableAccessor.tableExists(connection, tableName); -444 } -445}); -446 } -447 -448 @Override -449 public HTableDescriptor[] listTables() throws IOException { -450return listTables((Pattern)null, false); -451 } -452 -453 @Override -454 public HTableDescriptor[] listTables(Pattern pattern) throws IOException { -455
[23/51] [partial] hbase-site git commit: Published site at 9101fc246f86445006bfbcdfda5cc495016dc280.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/65565d77/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html index fea2b5a..c7a6cc4 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html @@ -1354,816 +1354,824 @@ 1346 */ 1347 public static void putsToMetaTable(final Connection connection, final ListPut ps) 1348 throws IOException { -1349try (Table t = getMetaHTable(connection)) { -1350 debugLogMutations(ps); -1351 t.put(ps); -1352} -1353 } -1354 -1355 /** -1356 * Delete the passed coded/code from the codehbase:meta/code table. -1357 * @param connection connection we're using -1358 * @param d Delete to add to hbase:meta -1359 */ -1360 private static void deleteFromMetaTable(final Connection connection, final Delete d) -1361 throws IOException { -1362ListDelete dels = new ArrayList(1); -1363dels.add(d); -1364deleteFromMetaTable(connection, dels); -1365 } -1366 -1367 /** -1368 * Delete the passed codedeletes/code from the codehbase:meta/code table. -1369 * @param connection connection we're using -1370 * @param deletes Deletes to add to hbase:meta This list should support #remove. -1371 */ -1372 private static void deleteFromMetaTable(final Connection connection, final ListDelete deletes) -1373 throws IOException { -1374try (Table t = getMetaHTable(connection)) { -1375 debugLogMutations(deletes); -1376 t.delete(deletes); -1377} -1378 } -1379 -1380 /** -1381 * Deletes some replica columns corresponding to replicas for the passed rows -1382 * @param metaRows rows in hbase:meta -1383 * @param replicaIndexToDeleteFrom the replica ID we would start deleting from -1384 * @param numReplicasToRemove how many replicas to remove -1385 * @param connection connection we're using to access meta table -1386 */ -1387 public static void removeRegionReplicasFromMeta(Setbyte[] metaRows, -1388int replicaIndexToDeleteFrom, int numReplicasToRemove, Connection connection) -1389 throws IOException { -1390int absoluteIndex = replicaIndexToDeleteFrom + numReplicasToRemove; -1391for (byte[] row : metaRows) { -1392 long now = EnvironmentEdgeManager.currentTime(); -1393 Delete deleteReplicaLocations = new Delete(row); -1394 for (int i = replicaIndexToDeleteFrom; i absoluteIndex; i++) { -1395 deleteReplicaLocations.addColumns(getCatalogFamily(), -1396 getServerColumn(i), now); -1397 deleteReplicaLocations.addColumns(getCatalogFamily(), -1398 getSeqNumColumn(i), now); -1399 deleteReplicaLocations.addColumns(getCatalogFamily(), -1400 getStartCodeColumn(i), now); -1401 } -1402 deleteFromMetaTable(connection, deleteReplicaLocations); -1403} -1404 } -1405 -1406 /** -1407 * Execute the passed codemutations/code against codehbase:meta/code table. -1408 * @param connection connection we're using -1409 * @param mutations Puts and Deletes to execute on hbase:meta -1410 * @throws IOException -1411 */ -1412 public static void mutateMetaTable(final Connection connection, -1413 final ListMutation mutations) -1414throws IOException { -1415Table t = getMetaHTable(connection); -1416try { -1417 debugLogMutations(mutations); -1418 t.batch(mutations, null); -1419} catch (InterruptedException e) { -1420 InterruptedIOException ie = new InterruptedIOException(e.getMessage()); -1421 ie.initCause(e); -1422 throw ie; -1423} finally { -1424 t.close(); -1425} -1426 } -1427 -1428 private static void addRegionStateToPut(Put put, RegionState.State state) throws IOException { -1429 put.add(CellBuilderFactory.create(CellBuilderType.SHALLOW_COPY) -1430.setRow(put.getRow()) -1431 .setFamily(HConstants.CATALOG_FAMILY) -1432 .setQualifier(getRegionStateColumn()) -1433 .setTimestamp(put.getTimestamp()) -1434.setType(Cell.Type.Put) -1435 .setValue(Bytes.toBytes(state.name())) -1436.build()); -1437 } -1438 -1439 /** -1440 * Adds daughter region infos to hbase:meta row for the specified region. Note that this does not -1441 * add its daughter's as different rows, but adds information about the daughters in the same row -1442 * as the parent. Use -1443 * {@link #splitRegion(Connection, RegionInfo, long, RegionInfo, RegionInfo, ServerName, int)} -1444 * if you want to do that. -1445 * @param connection connection we're using -1446 *
[23/51] [partial] hbase-site git commit: Published site at 0b28155d274910b4e667b949d51f78809a1eff0b.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e11cf2cb/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html index 42d0637..eb16038 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html @@ -80,21 +80,21 @@ 072import org.apache.hadoop.hbase.PleaseHoldException; 073import org.apache.hadoop.hbase.ReplicationPeerNotFoundException; 074import org.apache.hadoop.hbase.ScheduledChore; -075import org.apache.hadoop.hbase.ServerMetricsBuilder; -076import org.apache.hadoop.hbase.ServerName; -077import org.apache.hadoop.hbase.TableDescriptors; -078import org.apache.hadoop.hbase.TableName; -079import org.apache.hadoop.hbase.TableNotDisabledException; -080import org.apache.hadoop.hbase.TableNotFoundException; -081import org.apache.hadoop.hbase.UnknownRegionException; -082import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor; -083import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder; -084import org.apache.hadoop.hbase.client.MasterSwitchType; -085import org.apache.hadoop.hbase.client.RegionInfo; -086import org.apache.hadoop.hbase.client.Result; -087import org.apache.hadoop.hbase.client.TableDescriptor; -088import org.apache.hadoop.hbase.client.TableDescriptorBuilder; -089import org.apache.hadoop.hbase.client.TableState; +075import org.apache.hadoop.hbase.ServerName; +076import org.apache.hadoop.hbase.TableDescriptors; +077import org.apache.hadoop.hbase.TableName; +078import org.apache.hadoop.hbase.TableNotDisabledException; +079import org.apache.hadoop.hbase.TableNotFoundException; +080import org.apache.hadoop.hbase.UnknownRegionException; +081import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor; +082import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder; +083import org.apache.hadoop.hbase.client.MasterSwitchType; +084import org.apache.hadoop.hbase.client.RegionInfo; +085import org.apache.hadoop.hbase.client.Result; +086import org.apache.hadoop.hbase.client.TableDescriptor; +087import org.apache.hadoop.hbase.client.TableDescriptorBuilder; +088import org.apache.hadoop.hbase.client.TableState; +089import org.apache.hadoop.hbase.client.VersionInfoUtil; 090import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; 091import org.apache.hadoop.hbase.exceptions.DeserializationException; 092import org.apache.hadoop.hbase.exceptions.MergeRegionException; @@ -220,3477 +220,3481 @@ 212 213import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; 214import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState; -215import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionServerInfo; -216import org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas; -217import org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceViolationPolicy; -218import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription; -219import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos; -220 -221/** -222 * HMaster is the "master server" for HBase. An HBase cluster has one active -223 * master. If many masters are started, all compete. Whichever wins goes on to -224 * run the cluster. All others park themselves in their constructor until -225 * master or cluster shutdown or until the active master loses its lease in -226 * zookeeper. Thereafter, all running master jostle to take over master role. -227 * -228 * pThe Master can be asked shutdown the cluster. See {@link #shutdown()}. In -229 * this case it will tell all regionservers to go down and then wait on them -230 * all reporting in that they are down. This master will then shut itself down. -231 * -232 * pYou can also shutdown just this master. Call {@link #stopMaster()}. -233 * -234 * @see org.apache.zookeeper.Watcher -235 */ -236@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -237@SuppressWarnings("deprecation") -238public class HMaster extends HRegionServer implements MasterServices { -239 private static Logger LOG = LoggerFactory.getLogger(HMaster.class.getName()); -240 -241 /** -242 * Protection against zombie master. Started once Master accepts active responsibility and -243 * starts taking over responsibilities. Allows a finite time window before giving up ownership. -244 */ -245 private static class InitializationMonitor extends HasThread { -246/** The amount of time in milliseconds to sleep before checking initialization status. */ -247public static final String TIMEOUT_KEY =
[23/51] [partial] hbase-site git commit: Published site at 7d3750bd9fc9747623549c242cc4171e224b3eaf.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/3469cbc0/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html index fe1e077..90c31f4 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html @@ -1072,894 +1072,913 @@ 1064 1065 protected boolean waitServerReportEvent(final ServerName serverName, final Procedure proc) { 1066final ServerStateNode serverNode = regionStates.getOrCreateServer(serverName); -1067return serverNode.getReportEvent().suspendIfNotReady(proc); -1068 } -1069 -1070 protected void wakeServerReportEvent(final ServerStateNode serverNode) { -1071 serverNode.getReportEvent().wake(getProcedureScheduler()); -1072 } -1073 -1074 // -1075 // RIT chore -1076 // -1077 private static class RegionInTransitionChore extends ProcedureInMemoryChoreMasterProcedureEnv { -1078public RegionInTransitionChore(final int timeoutMsec) { -1079 super(timeoutMsec); -1080} -1081 -1082@Override -1083protected void periodicExecute(final MasterProcedureEnv env) { -1084 final AssignmentManager am = env.getAssignmentManager(); -1085 -1086 final RegionInTransitionStat ritStat = am.computeRegionInTransitionStat(); -1087 if (ritStat.hasRegionsOverThreshold()) { -1088for (RegionState hri: ritStat.getRegionOverThreshold()) { -1089 am.handleRegionOverStuckWarningThreshold(hri.getRegion()); -1090} -1091 } -1092 -1093 // update metrics -1094 am.updateRegionsInTransitionMetrics(ritStat); -1095} -1096 } -1097 -1098 public RegionInTransitionStat computeRegionInTransitionStat() { -1099final RegionInTransitionStat rit = new RegionInTransitionStat(getConfiguration()); -1100rit.update(this); -1101return rit; -1102 } -1103 -1104 public static class RegionInTransitionStat { -1105private final int ritThreshold; +1067if (serverNode == null) { +1068 LOG.warn("serverName=null; {}", proc); +1069} +1070return serverNode.getReportEvent().suspendIfNotReady(proc); +1071 } +1072 +1073 protected void wakeServerReportEvent(final ServerStateNode serverNode) { +1074 serverNode.getReportEvent().wake(getProcedureScheduler()); +1075 } +1076 +1077 // +1078 // RIT chore +1079 // +1080 private static class RegionInTransitionChore extends ProcedureInMemoryChoreMasterProcedureEnv { +1081public RegionInTransitionChore(final int timeoutMsec) { +1082 super(timeoutMsec); +1083} +1084 +1085@Override +1086protected void periodicExecute(final MasterProcedureEnv env) { +1087 final AssignmentManager am = env.getAssignmentManager(); +1088 +1089 final RegionInTransitionStat ritStat = am.computeRegionInTransitionStat(); +1090 if (ritStat.hasRegionsOverThreshold()) { +1091for (RegionState hri: ritStat.getRegionOverThreshold()) { +1092 am.handleRegionOverStuckWarningThreshold(hri.getRegion()); +1093} +1094 } +1095 +1096 // update metrics +1097 am.updateRegionsInTransitionMetrics(ritStat); +1098} +1099 } +1100 +1101 public RegionInTransitionStat computeRegionInTransitionStat() { +1102final RegionInTransitionStat rit = new RegionInTransitionStat(getConfiguration()); +1103rit.update(this); +1104return rit; +1105 } 1106 -1107private HashMapString, RegionState ritsOverThreshold = null; -1108private long statTimestamp; -1109private long oldestRITTime = 0; -1110private int totalRITsTwiceThreshold = 0; -private int totalRITs = 0; -1112 -1113@VisibleForTesting -1114public RegionInTransitionStat(final Configuration conf) { -1115 this.ritThreshold = -1116 conf.getInt(METRICS_RIT_STUCK_WARNING_THRESHOLD, DEFAULT_RIT_STUCK_WARNING_THRESHOLD); -1117} -1118 -1119public int getRITThreshold() { -1120 return ritThreshold; -1121} -1122 -1123public long getTimestamp() { -1124 return statTimestamp; -1125} -1126 -1127public int getTotalRITs() { -1128 return totalRITs; -1129} -1130 -1131public long getOldestRITTime() {
[23/51] [partial] hbase-site git commit: Published site at 997747076d8ec0b4346d7cb99c4b0667a7c14905.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4df09ed9/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html index 3da432b..d30fa8f 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatchOperation.html @@ -928,7690 +928,7698 @@ 920 CollectionHStore stores = this.stores.values(); 921 try { 922// update the stores that we are replaying -923 stores.forEach(HStore::startReplayingFromWAL); -924// Recover any edits if available. -925maxSeqId = Math.max(maxSeqId, -926 replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, status)); -927// Make sure mvcc is up to max. -928this.mvcc.advanceTo(maxSeqId); -929 } finally { -930// update the stores that we are done replaying -931 stores.forEach(HStore::stopReplayingFromWAL); -932 } -933} -934this.lastReplayedOpenRegionSeqId = maxSeqId; -935 -936 this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this)); -937this.writestate.flushRequested = false; -938this.writestate.compacting.set(0); -939 -940if (this.writestate.writesEnabled) { -941 // Remove temporary data left over from old regions -942 status.setStatus("Cleaning up temporary data from old regions"); -943 fs.cleanupTempDir(); -944} -945 -946if (this.writestate.writesEnabled) { -947 status.setStatus("Cleaning up detritus from prior splits"); -948 // Get rid of any splits or merges that were lost in-progress. Clean out -949 // these directories here on open. We may be opening a region that was -950 // being split but we crashed in the middle of it all. -951 fs.cleanupAnySplitDetritus(); -952 fs.cleanupMergesDir(); -953} -954 -955// Initialize split policy -956this.splitPolicy = RegionSplitPolicy.create(this, conf); -957 -958// Initialize flush policy -959this.flushPolicy = FlushPolicyFactory.create(this, conf); -960 -961long lastFlushTime = EnvironmentEdgeManager.currentTime(); -962for (HStore store: stores.values()) { -963 this.lastStoreFlushTimeMap.put(store, lastFlushTime); -964} -965 -966// Use maximum of log sequenceid or that which was found in stores -967// (particularly if no recovered edits, seqid will be -1). -968long maxSeqIdFromFile = -969 WALSplitter.getMaxRegionSequenceId(fs.getFileSystem(), fs.getRegionDir()); -970long nextSeqId = Math.max(maxSeqId, maxSeqIdFromFile) + 1; -971if (writestate.writesEnabled) { -972 WALSplitter.writeRegionSequenceIdFile(fs.getFileSystem(), fs.getRegionDir(), nextSeqId - 1); -973} -974 -975LOG.info("Opened {}; next sequenceid={}", this.getRegionInfo().getShortNameToLog(), nextSeqId); -976 -977// A region can be reopened if failed a split; reset flags -978this.closing.set(false); -979this.closed.set(false); -980 -981if (coprocessorHost != null) { -982 status.setStatus("Running coprocessor post-open hooks"); -983 coprocessorHost.postOpen(); -984} +923LOG.debug("replaying wal for " + this.getRegionInfo().getEncodedName()); +924 stores.forEach(HStore::startReplayingFromWAL); +925// Recover any edits if available. +926maxSeqId = Math.max(maxSeqId, +927 replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, status)); +928// Make sure mvcc is up to max. +929this.mvcc.advanceTo(maxSeqId); +930 } finally { +931LOG.debug("stopping wal replay for " + this.getRegionInfo().getEncodedName()); +932// update the stores that we are done replaying +933 stores.forEach(HStore::stopReplayingFromWAL); +934 } +935} +936this.lastReplayedOpenRegionSeqId = maxSeqId; +937 +938 this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this)); +939this.writestate.flushRequested = false; +940this.writestate.compacting.set(0); +941 +942if (this.writestate.writesEnabled) { +943 LOG.debug("Cleaning up temporary data for " + this.getRegionInfo().getEncodedName()); +944 // Remove temporary data left over from old regions +945 status.setStatus("Cleaning up temporary data from old regions"); +946 fs.cleanupTempDir(); +947} +948 +949if (this.writestate.writesEnabled) { +950 status.setStatus("Cleaning up detritus from prior splits"); +951 // Get rid of any splits or merges that were lost in-progress. Clean out +952 // these
[23/51] [partial] hbase-site git commit: Published site at f3d1c021de2264301f68eadb9ef126ff83d7ef53.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/883dde2f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.PeriodicMemStoreFlusher.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.PeriodicMemStoreFlusher.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.PeriodicMemStoreFlusher.html index 9644187..b979909 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.PeriodicMemStoreFlusher.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.PeriodicMemStoreFlusher.html @@ -1199,2602 +1199,2599 @@ 1191ClusterStatusProtos.ServerLoad sl = buildServerLoad(reportStartTime, reportEndTime); 1192try { 1193 RegionServerReportRequest.Builder request = RegionServerReportRequest.newBuilder(); -1194 ServerName sn = ServerName.parseVersionedServerName(this.serverName.getVersionedBytes()); -1195 request.setServer(ProtobufUtil.toServerName(sn)); -1196 request.setLoad(sl); -1197 rss.regionServerReport(null, request.build()); -1198} catch (ServiceException se) { -1199 IOException ioe = ProtobufUtil.getRemoteException(se); -1200 if (ioe instanceof YouAreDeadException) { -1201// This will be caught and handled as a fatal error in run() -1202throw ioe; -1203 } -1204 if (rssStub == rss) { -1205rssStub = null; -1206 } -1207 // Couldn't connect to the master, get location from zk and reconnect -1208 // Method blocks until new master is found or we are stopped -1209 createRegionServerStatusStub(true); -1210} -1211 } -1212 -1213 /** -1214 * Reports the given map of Regions and their size on the filesystem to the active Master. -1215 * -1216 * @param regionSizeStore The store containing region sizes -1217 * @return false if FileSystemUtilizationChore should pause reporting to master. true otherwise -1218 */ -1219 public boolean reportRegionSizesForQuotas(RegionSizeStore regionSizeStore) { -1220 RegionServerStatusService.BlockingInterface rss = rssStub; -1221if (rss == null) { -1222 // the current server could be stopping. -1223 LOG.trace("Skipping Region size report to HMaster as stub is null"); -1224 return true; -1225} -1226try { -1227 buildReportAndSend(rss, regionSizeStore); -1228} catch (ServiceException se) { -1229 IOException ioe = ProtobufUtil.getRemoteException(se); -1230 if (ioe instanceof PleaseHoldException) { -1231LOG.trace("Failed to report region sizes to Master because it is initializing." -1232+ " This will be retried.", ioe); -1233// The Master is coming up. Will retry the report later. Avoid re-creating the stub. -1234return true; -1235 } -1236 if (rssStub == rss) { -1237rssStub = null; -1238 } -1239 createRegionServerStatusStub(true); -1240 if (ioe instanceof DoNotRetryIOException) { -1241DoNotRetryIOException doNotRetryEx = (DoNotRetryIOException) ioe; -1242if (doNotRetryEx.getCause() != null) { -1243 Throwable t = doNotRetryEx.getCause(); -1244 if (t instanceof UnsupportedOperationException) { -1245LOG.debug("master doesn't support ReportRegionSpaceUse, pause before retrying"); -1246return false; -1247 } -1248} -1249 } -1250 LOG.debug("Failed to report region sizes to Master. This will be retried.", ioe); -1251} -1252return true; -1253 } -1254 -1255 /** -1256 * Builds the region size report and sends it to the master. Upon successful sending of the -1257 * report, the region sizes that were sent are marked as sent. -1258 * -1259 * @param rss The stub to send to the Master -1260 * @param regionSizeStore The store containing region sizes -1261 */ -1262 void buildReportAndSend(RegionServerStatusService.BlockingInterface rss, -1263 RegionSizeStore regionSizeStore) throws ServiceException { -1264RegionSpaceUseReportRequest request = -1265 buildRegionSpaceUseReportRequest(Objects.requireNonNull(regionSizeStore)); -1266rss.reportRegionSpaceUse(null, request); -1267// Record the number of size reports sent -1268if (metricsRegionServer != null) { -1269 metricsRegionServer.incrementNumRegionSizeReportsSent(regionSizeStore.size()); -1270} -1271 } -1272 -1273 /** -1274 * Builds a {@link RegionSpaceUseReportRequest} protobuf message from the region size map. -1275 * -1276 * @param regionSizeStore The size in bytes of regions -1277 * @return The corresponding protocol buffer message. -1278 */ -1279 RegionSpaceUseReportRequest buildRegionSpaceUseReportRequest(RegionSizeStore regionSizes) { -1280RegionSpaceUseReportRequest.Builder request =
[23/51] [partial] hbase-site git commit: Published site at cf529f18a9959589fa635f78df4840472526ea2c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7bcc960d/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html index 8b2674f..274eb54 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html @@ -262,1413 +262,1417 @@ 254}); 255 } 256 -257 public void preModifyNamespace(final NamespaceDescriptor ns) throws IOException { -258 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { -259 @Override -260 public void call(MasterObserver observer) throws IOException { -261observer.preModifyNamespace(this, ns); -262 } -263}); -264 } -265 -266 public void postModifyNamespace(final NamespaceDescriptor ns) throws IOException { -267 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { -268 @Override -269 public void call(MasterObserver observer) throws IOException { -270 observer.postModifyNamespace(this, ns); -271 } -272}); -273 } -274 -275 public void preGetNamespaceDescriptor(final String namespaceName) -276 throws IOException { -277 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { -278 @Override -279 public void call(MasterObserver observer) throws IOException { -280 observer.preGetNamespaceDescriptor(this, namespaceName); -281 } -282}); -283 } -284 -285 public void postGetNamespaceDescriptor(final NamespaceDescriptor ns) -286 throws IOException { -287 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { -288 @Override -289 public void call(MasterObserver observer) throws IOException { -290 observer.postGetNamespaceDescriptor(this, ns); -291 } -292}); -293 } -294 -295 public void preListNamespaceDescriptors(final ListNamespaceDescriptor descriptors) -296 throws IOException { -297 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { -298 @Override -299 public void call(MasterObserver observer) throws IOException { -300 observer.preListNamespaceDescriptors(this, descriptors); -301 } -302}); -303 } -304 -305 public void postListNamespaceDescriptors(final ListNamespaceDescriptor descriptors) -306 throws IOException { -307 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { -308 @Override -309 public void call(MasterObserver observer) throws IOException { -310 observer.postListNamespaceDescriptors(this, descriptors); -311 } -312}); -313 } -314 -315 /* Implementation of hooks for invoking MasterObservers */ +257 public void preModifyNamespace(final NamespaceDescriptor currentNsDescriptor, +258final NamespaceDescriptor newNsDescriptor) throws IOException { +259 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { +260 @Override +261 public void call(MasterObserver observer) throws IOException { +262observer.preModifyNamespace(this, currentNsDescriptor, newNsDescriptor); +263 } +264}); +265 } +266 +267 public void postModifyNamespace(final NamespaceDescriptor oldNsDescriptor, +268final NamespaceDescriptor currentNsDescriptor) throws IOException { +269 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { +270 @Override +271 public void call(MasterObserver observer) throws IOException { +272 observer.postModifyNamespace(this, oldNsDescriptor, currentNsDescriptor); +273 } +274}); +275 } +276 +277 public void preGetNamespaceDescriptor(final String namespaceName) +278 throws IOException { +279 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { +280 @Override +281 public void call(MasterObserver observer) throws IOException { +282 observer.preGetNamespaceDescriptor(this, namespaceName); +283 } +284}); +285 } +286 +287 public void postGetNamespaceDescriptor(final NamespaceDescriptor ns) +288 throws IOException { +289 execOperation(coprocEnvironments.isEmpty() ? null : new MasterObserverOperation() { +290 @Override +291 public void call(MasterObserver observer) throws IOException { +292 observer.postGetNamespaceDescriptor(this, ns); +293 } +294}); +295 } +296 +297 public void preListNamespaceDescriptors(final ListNamespaceDescriptor descriptors) +298 throws IOException { +299 execOperation(coprocEnvironments.isEmpty() ? null : new
[23/51] [partial] hbase-site git commit: Published site at 021f66d11d2cbb7308308093e29e69d6e7661ee9.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/92a26cfb/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html index 4a879bb..7d27402 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html @@ -300,7 +300,7 @@ 292 private MapString, com.google.protobuf.Service coprocessorServiceHandlers = Maps.newHashMap(); 293 294 // Track data size in all memstores -295 private final MemStoreSizing memStoreSize = new MemStoreSizing(); +295 private final MemStoreSizing memStoreSizing = new ThreadSafeMemStoreSizing(); 296 private final RegionServicesForStores regionServicesForStores = new RegionServicesForStores(this); 297 298 // Debug possible data loss due to WAL off @@ -1218,7389 +1218,7399 @@ 1210 * Increase the size of mem store in this region and the size of global mem 1211 * store 1212 */ -1213 public void incMemStoreSize(MemStoreSize memStoreSize) { -1214if (this.rsAccounting != null) { -1215 rsAccounting.incGlobalMemStoreSize(memStoreSize); -1216} -1217long dataSize; -1218synchronized (this.memStoreSize) { -1219 this.memStoreSize.incMemStoreSize(memStoreSize); -1220 dataSize = this.memStoreSize.getDataSize(); -1221} -1222 checkNegativeMemStoreDataSize(dataSize, memStoreSize.getDataSize()); -1223 } -1224 -1225 public void decrMemStoreSize(MemStoreSize memStoreSize) { -1226if (this.rsAccounting != null) { -1227 rsAccounting.decGlobalMemStoreSize(memStoreSize); -1228} -1229long size; -1230synchronized (this.memStoreSize) { -1231 this.memStoreSize.decMemStoreSize(memStoreSize); -1232 size = this.memStoreSize.getDataSize(); +1213 void incMemStoreSize(MemStoreSize mss) { +1214incMemStoreSize(mss.getDataSize(), mss.getHeapSize(), mss.getOffHeapSize()); +1215 } +1216 +1217 void incMemStoreSize(long dataSizeDelta, long heapSizeDelta, long offHeapSizeDelta) { +1218if (this.rsAccounting != null) { +1219 rsAccounting.incGlobalMemStoreSize(dataSizeDelta, heapSizeDelta, offHeapSizeDelta); +1220} +1221long dataSize = +1222 this.memStoreSizing.incMemStoreSize(dataSizeDelta, heapSizeDelta, offHeapSizeDelta); +1223 checkNegativeMemStoreDataSize(dataSize, dataSizeDelta); +1224 } +1225 +1226 void decrMemStoreSize(MemStoreSize mss) { +1227decrMemStoreSize(mss.getDataSize(), mss.getHeapSize(), mss.getOffHeapSize()); +1228 } +1229 +1230 void decrMemStoreSize(long dataSizeDelta, long heapSizeDelta, long offHeapSizeDelta) { +1231if (this.rsAccounting != null) { +1232 rsAccounting.decGlobalMemStoreSize(dataSizeDelta, heapSizeDelta, offHeapSizeDelta); 1233} -1234checkNegativeMemStoreDataSize(size, -memStoreSize.getDataSize()); -1235 } -1236 -1237 private void checkNegativeMemStoreDataSize(long memStoreDataSize, long delta) { -1238// This is extremely bad if we make memStoreSize negative. Log as much info on the offending -1239// caller as possible. (memStoreSize might be a negative value already -- freeing memory) -1240if (memStoreDataSize 0) { -1241 LOG.error("Asked to modify this region's (" + this.toString() -1242 + ") memStoreSize to a negative value which is incorrect. Current memStoreSize=" -1243 + (memStoreDataSize - delta) + ", delta=" + delta, new Exception()); -1244} -1245 } -1246 -1247 @Override -1248 public RegionInfo getRegionInfo() { -1249return this.fs.getRegionInfo(); -1250 } -1251 -1252 /** -1253 * @return Instance of {@link RegionServerServices} used by this HRegion. -1254 * Can be null. -1255 */ -1256 RegionServerServices getRegionServerServices() { -1257return this.rsServices; -1258 } -1259 -1260 @Override -1261 public long getReadRequestsCount() { -1262return readRequestsCount.sum(); -1263 } -1264 -1265 @Override -1266 public long getFilteredReadRequestsCount() { -1267return filteredReadRequestsCount.sum(); -1268 } -1269 -1270 @Override -1271 public long getWriteRequestsCount() { -1272return writeRequestsCount.sum(); -1273 } -1274 -1275 @Override -1276 public long getMemStoreDataSize() { -1277return memStoreSize.getDataSize(); -1278 } -1279 -1280 @Override -1281 public long getMemStoreHeapSize() { -1282return memStoreSize.getHeapSize(); -1283 } -1284 -1285 @Override -1286 public long getMemStoreOffHeapSize() { -1287return memStoreSize.getOffHeapSize(); -1288 } -1289 -1290 /** @return store services for this region, to access services required by store level needs */ -1291 public RegionServicesForStores
[23/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CASTableTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CASTableTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CASTableTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CASTableTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CASTableTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[23/51] [partial] hbase-site git commit: Published site at 87f5b5f3411d96c31b4cb61b9a57ced22be91d1f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/de18d468/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.RegionRepairException.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.RegionRepairException.html b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.RegionRepairException.html index e1bc325..63e7421 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.RegionRepairException.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.RegionRepairException.html @@ -66,5125 +66,5224 @@ 058import java.util.concurrent.TimeoutException; 059import java.util.concurrent.atomic.AtomicBoolean; 060import java.util.concurrent.atomic.AtomicInteger; -061import org.apache.commons.io.IOUtils; -062import org.apache.commons.lang3.RandomStringUtils; -063import org.apache.commons.lang3.StringUtils; -064import org.apache.hadoop.conf.Configuration; -065import org.apache.hadoop.conf.Configured; -066import org.apache.hadoop.fs.FSDataOutputStream; -067import org.apache.hadoop.fs.FileStatus; -068import org.apache.hadoop.fs.FileSystem; -069import org.apache.hadoop.fs.Path; -070import org.apache.hadoop.fs.permission.FsAction; -071import org.apache.hadoop.fs.permission.FsPermission; -072import org.apache.hadoop.hbase.Abortable; -073import org.apache.hadoop.hbase.Cell; -074import org.apache.hadoop.hbase.CellUtil; -075import org.apache.hadoop.hbase.ClusterMetrics; -076import org.apache.hadoop.hbase.ClusterMetrics.Option; -077import org.apache.hadoop.hbase.HBaseConfiguration; -078import org.apache.hadoop.hbase.HBaseInterfaceAudience; -079import org.apache.hadoop.hbase.HConstants; -080import org.apache.hadoop.hbase.HRegionInfo; -081import org.apache.hadoop.hbase.HRegionLocation; -082import org.apache.hadoop.hbase.KeyValue; -083import org.apache.hadoop.hbase.MasterNotRunningException; -084import org.apache.hadoop.hbase.MetaTableAccessor; -085import org.apache.hadoop.hbase.RegionLocations; -086import org.apache.hadoop.hbase.ServerName; -087import org.apache.hadoop.hbase.TableName; -088import org.apache.hadoop.hbase.ZooKeeperConnectionException; -089import org.apache.hadoop.hbase.client.Admin; -090import org.apache.hadoop.hbase.client.ClusterConnection; -091import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor; -092import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder; -093import org.apache.hadoop.hbase.client.Connection; -094import org.apache.hadoop.hbase.client.ConnectionFactory; -095import org.apache.hadoop.hbase.client.Delete; -096import org.apache.hadoop.hbase.client.Get; -097import org.apache.hadoop.hbase.client.Put; -098import org.apache.hadoop.hbase.client.RegionInfo; -099import org.apache.hadoop.hbase.client.RegionInfoBuilder; -100import org.apache.hadoop.hbase.client.RegionReplicaUtil; -101import org.apache.hadoop.hbase.client.Result; -102import org.apache.hadoop.hbase.client.RowMutations; -103import org.apache.hadoop.hbase.client.Table; -104import org.apache.hadoop.hbase.client.TableDescriptor; -105import org.apache.hadoop.hbase.client.TableDescriptorBuilder; -106import org.apache.hadoop.hbase.client.TableState; -107import org.apache.hadoop.hbase.io.FileLink; -108import org.apache.hadoop.hbase.io.HFileLink; -109import org.apache.hadoop.hbase.io.hfile.CacheConfig; -110import org.apache.hadoop.hbase.io.hfile.HFile; -111import org.apache.hadoop.hbase.log.HBaseMarkers; -112import org.apache.hadoop.hbase.master.MasterFileSystem; -113import org.apache.hadoop.hbase.master.RegionState; -114import org.apache.hadoop.hbase.regionserver.HRegion; -115import org.apache.hadoop.hbase.regionserver.HRegionFileSystem; -116import org.apache.hadoop.hbase.regionserver.StoreFileInfo; -117import org.apache.hadoop.hbase.replication.ReplicationException; -118import org.apache.hadoop.hbase.security.AccessDeniedException; -119import org.apache.hadoop.hbase.security.UserProvider; -120import org.apache.hadoop.hbase.util.Bytes.ByteArrayComparator; -121import org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE; -122import org.apache.hadoop.hbase.util.hbck.HFileCorruptionChecker; -123import org.apache.hadoop.hbase.util.hbck.ReplicationChecker; -124import org.apache.hadoop.hbase.util.hbck.TableIntegrityErrorHandler; -125import org.apache.hadoop.hbase.util.hbck.TableIntegrityErrorHandlerImpl; -126import org.apache.hadoop.hbase.wal.WAL; -127import org.apache.hadoop.hbase.wal.WALFactory; -128import org.apache.hadoop.hbase.wal.WALSplitter; -129import org.apache.hadoop.hbase.zookeeper.MetaTableLocator; -130import org.apache.hadoop.hbase.zookeeper.ZKUtil; -131import org.apache.hadoop.hbase.zookeeper.ZKWatcher; -132import org.apache.hadoop.hbase.zookeeper.ZNodePaths; -133import org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException; -134import org.apache.hadoop.ipc.RemoteException; -135import
[23/51] [partial] hbase-site git commit: Published site at 2912c953551bedbfbf30c32c156ed7bb187d54c3.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d220bc5e/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.SyncMetrics.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.SyncMetrics.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.SyncMetrics.html index e6e43ee..a8b77ae 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.SyncMetrics.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.SyncMetrics.html @@ -367,650 +367,650 @@ 359lock.lock(); 360try { 361 LOG.trace("Starting WAL Procedure Store lease recovery"); -362 FileStatus[] oldLogs = getLogFiles(); -363 while (isRunning()) { +362 while (isRunning()) { +363FileStatus[] oldLogs = getLogFiles(); 364// Get Log-MaxID and recover lease on old logs 365try { 366 flushLogId = initOldLogs(oldLogs); 367} catch (FileNotFoundException e) { 368 LOG.warn("Someone else is active and deleted logs. retrying.", e); -369 oldLogs = getLogFiles(); -370 continue; -371} -372 -373// Create new state-log -374if (!rollWriter(flushLogId + 1)) { -375 // someone else has already created this log -376 LOG.debug("Someone else has already created log " + flushLogId); -377 continue; -378} -379 -380// We have the lease on the log -381oldLogs = getLogFiles(); -382if (getMaxLogId(oldLogs) flushLogId) { -383 if (LOG.isDebugEnabled()) { -384LOG.debug("Someone else created new logs. Expected maxLogId " + flushLogId); -385 } -386 logs.getLast().removeFile(this.walArchiveDir); -387 continue; -388} -389 -390LOG.trace("Lease acquired for flushLogId={}", flushLogId); -391break; -392 } -393} finally { -394 lock.unlock(); -395} -396 } -397 -398 @Override -399 public void load(final ProcedureLoader loader) throws IOException { -400lock.lock(); -401try { -402 if (logs.isEmpty()) { -403throw new RuntimeException("recoverLease() must be called before loading data"); -404 } -405 -406 // Nothing to do, If we have only the current log. -407 if (logs.size() == 1) { -408LOG.trace("No state logs to replay."); -409loader.setMaxProcId(0); -410return; -411 } -412 -413 // Load the old logs -414 final IteratorProcedureWALFile it = logs.descendingIterator(); -415 it.next(); // Skip the current log -416 -417 ProcedureWALFormat.load(it, storeTracker, new ProcedureWALFormat.Loader() { -418@Override -419public void setMaxProcId(long maxProcId) { -420 loader.setMaxProcId(maxProcId); -421} -422 -423@Override -424public void load(ProcedureIterator procIter) throws IOException { -425 loader.load(procIter); -426} -427 -428@Override -429public void handleCorrupted(ProcedureIterator procIter) throws IOException { -430 loader.handleCorrupted(procIter); -431} -432 -433@Override -434public void markCorruptedWAL(ProcedureWALFile log, IOException e) { -435 if (corruptedLogs == null) { -436corruptedLogs = new HashSet(); -437 } -438 corruptedLogs.add(log); -439 // TODO: sideline corrupted log -440} -441 }); -442} finally { -443 try { -444// try to cleanup inactive wals and complete the operation -445buildHoldingCleanupTracker(); -446tryCleanupLogsOnLoad(); -447loading.set(false); -448 } finally { -449lock.unlock(); -450 } -451} -452 } -453 -454 private void tryCleanupLogsOnLoad() { -455// nothing to cleanup. -456if (logs.size() = 1) return; -457 -458// the config says to not cleanup wals on load. -459if (!conf.getBoolean(EXEC_WAL_CLEANUP_ON_LOAD_CONF_KEY, -460 DEFAULT_EXEC_WAL_CLEANUP_ON_LOAD_CONF_KEY)) { -461 LOG.debug("WALs cleanup on load is not enabled: " + getActiveLogs()); -462 return; -463} -464 -465try { -466 periodicRoll(); -467} catch (IOException e) { -468 LOG.warn("Unable to cleanup logs on load: " + e.getMessage(), e); -469} -470 } -471 -472 @Override -473 public void insert(final Procedure proc, final Procedure[] subprocs) { -474if (LOG.isTraceEnabled()) { -475 LOG.trace("Insert " + proc + ", subproc=" + Arrays.toString(subprocs)); -476} -477 -478ByteSlot slot = acquireSlot(); -479try { -480 // Serialize the insert -481 long[] subProcIds = null; -482 if (subprocs != null) { -483
[23/51] [partial] hbase-site git commit: Published site at 2a2258656b2fcd92b967131b6c1f037363553bc4.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/src-html/org/apache/hadoop/hbase/CompoundConfiguration.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/CompoundConfiguration.html b/devapidocs/src-html/org/apache/hadoop/hbase/CompoundConfiguration.html index 1038256..89735e7 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/CompoundConfiguration.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/CompoundConfiguration.html @@ -36,398 +36,399 @@ 028import java.util.List; 029import java.util.Map; 030 -031import org.apache.commons.collections4.iterators.UnmodifiableIterator; -032import org.apache.hadoop.conf.Configuration; -033import org.apache.hadoop.hbase.util.Bytes; -034import org.apache.yetus.audience.InterfaceAudience; -035 -036/** -037 * Do a shallow merge of multiple KV configuration pools. This is a very useful -038 * utility class to easily add per-object configurations in addition to wider -039 * scope settings. This is different from Configuration.addResource() -040 * functionality, which performs a deep merge and mutates the common data -041 * structure. -042 * p -043 * The iterator on CompoundConfiguration is unmodifiable. Obtaining iterator is an expensive -044 * operation. -045 * p -046 * For clarity: the shallow merge allows the user to mutate either of the -047 * configuration objects and have changes reflected everywhere. In contrast to a -048 * deep merge, that requires you to explicitly know all applicable copies to -049 * propagate changes. -050 * -051 * WARNING: The values set in the CompoundConfiguration are do not handle Property variable -052 * substitution. However, if they are set in the underlying configuration substitutions are -053 * done. -054 */ -055@InterfaceAudience.Private -056public class CompoundConfiguration extends Configuration { -057 -058 private Configuration mutableConf = null; -059 -060 /** -061 * Default Constructor. Initializes empty configuration -062 */ -063 public CompoundConfiguration() { -064 } -065 -066 // Devs: these APIs are the same contract as their counterparts in -067 // Configuration.java -068 private interface ImmutableConfigMap extends IterableMap.EntryString,String { -069String get(String key); -070String getRaw(String key); -071Class? getClassByName(String name) throws ClassNotFoundException; -072int size(); -073 } -074 -075 private final ListImmutableConfigMap configs = new ArrayList(); -076 -077 static class ImmutableConfWrapper implements ImmutableConfigMap { -078 private final Configuration c; -079 -080ImmutableConfWrapper(Configuration conf) { -081 c = conf; -082} -083 -084@Override -085public IteratorMap.EntryString,String iterator() { -086 return c.iterator(); -087} -088 -089@Override -090public String get(String key) { -091 return c.get(key); -092} -093 -094@Override -095public String getRaw(String key) { -096 return c.getRaw(key); -097} -098 -099@Override -100public Class? getClassByName(String name) -101throws ClassNotFoundException { -102 return c.getClassByName(name); -103} -104 -105@Override -106public int size() { -107 return c.size(); -108} -109 -110@Override -111public String toString() { -112 return c.toString(); -113} -114 } -115 -116 /** -117 * If set has been called, it will create a mutableConf. This converts the mutableConf to an -118 * immutable one and resets it to allow a new mutable conf. This is used when a new map or -119 * conf is added to the compound configuration to preserve proper override semantics. -120 */ -121 void freezeMutableConf() { -122if (mutableConf == null) { -123 // do nothing if there is no current mutableConf -124 return; -125} -126 -127this.configs.add(0, new ImmutableConfWrapper(mutableConf)); -128mutableConf = null; -129 } -130 -131 /** -132 * Add Hadoop Configuration object to config list. -133 * The added configuration overrides the previous ones if there are name collisions. -134 * @param conf configuration object -135 * @return this, for builder pattern -136 */ -137 public CompoundConfiguration add(final Configuration conf) { -138freezeMutableConf(); -139 -140if (conf instanceof CompoundConfiguration) { -141 this.configs.addAll(0, ((CompoundConfiguration) conf).configs); -142 return this; -143} -144// put new config at the front of the list (top priority) -145this.configs.add(0, new ImmutableConfWrapper(conf)); -146return this; -147 } -148 -149 /** -150 * Add Bytes map to config list. This map is generally -151 * created by HTableDescriptor or HColumnDescriptor, but can be abstractly -152 * used. The added configuration overrides the previous ones if there are -153 * name
[23/51] [partial] hbase-site git commit: Published site at e468b4022f76688851b3e0c34722f01a56bd624f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/16541468/checkstyle.rss -- diff --git a/checkstyle.rss b/checkstyle.rss index 52f3dab..213814f 100644 --- a/checkstyle.rss +++ b/checkstyle.rss @@ -25,8 +25,8 @@ under the License. en-us 2007 - 2018 The Apache Software Foundation - File: 3592, - Errors: 16020, + File: 3593, + Errors: 15918, Warnings: 0, Infos: 0 @@ -1656,7 +1656,7 @@ under the License. - http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter.java;>org/apache/hadoop/hbase/filter/TestMultiRowRangeFilter.java + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.client.TestResult.java;>org/apache/hadoop/hbase/client/TestResult.java 0 @@ -1665,12 +1665,12 @@ under the License. 0 - 2 + 3 - http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.client.TestResult.java;>org/apache/hadoop/hbase/client/TestResult.java + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter.java;>org/apache/hadoop/hbase/filter/TestMultiRowRangeFilter.java 0 @@ -1679,7 +1679,7 @@ under the License. 0 - 3 + 2 @@ -1894,7 +1894,7 @@ under the License. - http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.mob.MobUtils.java;>org/apache/hadoop/hbase/mob/MobUtils.java + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.DateTieredStoreEngine.java;>org/apache/hadoop/hbase/regionserver/DateTieredStoreEngine.java 0 @@ -1903,12 +1903,12 @@ under the License. 0 - 15 + 1 - http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.DateTieredStoreEngine.java;>org/apache/hadoop/hbase/regionserver/DateTieredStoreEngine.java + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.mob.MobUtils.java;>org/apache/hadoop/hbase/mob/MobUtils.java 0 @@ -1917,7 +1917,7 @@ under the License. 0 - 1 + 15 @@ -2286,7 +2286,7 @@ under the License. - http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.coprocessor.TestCoprocessorMetrics.java;>org/apache/hadoop/hbase/coprocessor/TestCoprocessorMetrics.java + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.io.hfile.TestForceCacheImportantBlocks.java;>org/apache/hadoop/hbase/io/hfile/TestForceCacheImportantBlocks.java 0 @@ -2295,12 +2295,12 @@ under the License. 0 - 1 + 3 - http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.io.hfile.TestForceCacheImportantBlocks.java;>org/apache/hadoop/hbase/io/hfile/TestForceCacheImportantBlocks.java + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.coprocessor.TestCoprocessorMetrics.java;>org/apache/hadoop/hbase/coprocessor/TestCoprocessorMetrics.java 0 @@ -2309,7 +2309,7 @@ under the License. 0 - 3 + 1 @@ -3686,7 +3686,7 @@ under the License. - http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.RegionServerRunningException.java;>org/apache/hadoop/hbase/regionserver/RegionServerRunningException.java +
[23/51] [partial] hbase-site git commit: Published site at 64061f896fe21512504e3886a400759e88b519da.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f1ebf5b6/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html index c15c700..fa49042 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html @@ -511,13 +511,18 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. ServerNameserverName) +static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListPairhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">Long +MetaTableAccessor.getTableEncodedRegionNameAndLastBarrier(Connectionconn, + TableNametableName) + + static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo MetaTableAccessor.getTableRegions(Connectionconnection, TableNametableName) Gets all of the regions of the specified table. - + static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo MetaTableAccessor.getTableRegions(Connectionconnection, TableNametableName, @@ -525,14 +530,14 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. Gets all of the regions of the specified table. - + static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListPairRegionInfo,ServerName MetaTableAccessor.getTableRegionsAndLocations(Connectionconnection, TableNametableName) Do not use this method to get meta table regions, use methods in MetaTableLocator instead. - + static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListPairRegionInfo,ServerName MetaTableAccessor.getTableRegionsAndLocations(Connectionconnection, TableNametableName, @@ -540,20 +545,20 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. Do not use this method to get meta table regions, use methods in MetaTableLocator instead. - + static TableState MetaTableAccessor.getTableState(Connectionconn, TableNametableName) Fetch table state for given table from META table - + static https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapTableName,TableState MetaTableAccessor.getTableStates(Connectionconn) Fetch table states from META table - + static void MetaTableAccessor.mergeRegions(Connectionconnection, RegionInfomergedRegion, @@ -566,7 +571,7 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. Merge the two regions into one in an atomic operation. - + private static void MetaTableAccessor.multiMutate(Connectionconnection, Tabletable, @@ -575,21 +580,21 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. Performs an atomic multi-mutate operation against the given table. - + private static void MetaTableAccessor.multiMutate(Connectionconnection, Tabletable, byte[]row, Mutation...mutations) - + static void MetaTableAccessor.mutateMetaTable(Connectionconnection, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListMutationmutations) Execute the passed mutations against hbase:meta table. - + static void MetaTableAccessor.overwriteRegions(Connectionconnection, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInforegionInfos, @@ -597,21 +602,21 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. Overwrites the specified regions from hbase:meta. - + static void MetaTableAccessor.putsToMetaTable(Connectionconnection, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListPutps) Put the passed ps to the hbase:meta table. - + private static void MetaTableAccessor.putToMetaTable(Connectionconnection, Putp) Put the passed p to the hbase:meta table. - + static void
[23/51] [partial] hbase-site git commit: Published site at 4cb40e6d846ce1f28ffb40d388c9efb753197813.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4dc2a2e8/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html index d5a3666..d3a9f2d 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/ColumnFamilyDescriptor.html @@ -320,8 +320,10 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncHBaseAdmin.addColumnFamily(TableNametableName, - ColumnFamilyDescriptorcolumnFamily) +AsyncAdmin.addColumnFamily(TableNametableName, + ColumnFamilyDescriptorcolumnFamily) +Add a column family to an existing table. + void @@ -331,20 +333,18 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -void -HBaseAdmin.addColumnFamily(TableNametableName, +https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void +RawAsyncHBaseAdmin.addColumnFamily(TableNametableName, ColumnFamilyDescriptorcolumnFamily) -https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncAdmin.addColumnFamily(TableNametableName, - ColumnFamilyDescriptorcolumnFamily) -Add a column family to an existing table. - +void +HBaseAdmin.addColumnFamily(TableNametableName, + ColumnFamilyDescriptorcolumnFamily) https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -RawAsyncHBaseAdmin.addColumnFamily(TableNametableName, +AsyncHBaseAdmin.addColumnFamily(TableNametableName, ColumnFamilyDescriptorcolumnFamily) @@ -386,8 +386,10 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncHBaseAdmin.modifyColumnFamily(TableNametableName, - ColumnFamilyDescriptorcolumnFamily) +AsyncAdmin.modifyColumnFamily(TableNametableName, + ColumnFamilyDescriptorcolumnFamily) +Modify an existing column family on a table. + void @@ -397,20 +399,18 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -void -HBaseAdmin.modifyColumnFamily(TableNametableName, +https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void +RawAsyncHBaseAdmin.modifyColumnFamily(TableNametableName, ColumnFamilyDescriptorcolumnFamily) -https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void -AsyncAdmin.modifyColumnFamily(TableNametableName, - ColumnFamilyDescriptorcolumnFamily) -Modify an existing column family on a table. - +void +HBaseAdmin.modifyColumnFamily(TableNametableName, + ColumnFamilyDescriptorcolumnFamily) https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void
[23/51] [partial] hbase-site git commit: Published site at 8ab7b20f48951d77945181024f5e15842bc253c4.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6eb695c8/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html index ecf500c..0cd5a4e 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html @@ -238,8355 +238,8368 @@ 230 public static final String HBASE_MAX_CELL_SIZE_KEY = "hbase.server.keyvalue.maxsize"; 231 public static final int DEFAULT_MAX_CELL_SIZE = 10485760; 232 -233 public static final String HBASE_REGIONSERVER_MINIBATCH_SIZE = -234 "hbase.regionserver.minibatch.size"; -235 public static final int DEFAULT_HBASE_REGIONSERVER_MINIBATCH_SIZE = 2; -236 -237 /** -238 * This is the global default value for durability. All tables/mutations not -239 * defining a durability or using USE_DEFAULT will default to this value. -240 */ -241 private static final Durability DEFAULT_DURABILITY = Durability.SYNC_WAL; +233 /** +234 * This is the global default value for durability. All tables/mutations not +235 * defining a durability or using USE_DEFAULT will default to this value. +236 */ +237 private static final Durability DEFAULT_DURABILITY = Durability.SYNC_WAL; +238 +239 public static final String HBASE_REGIONSERVER_MINIBATCH_SIZE = +240 "hbase.regionserver.minibatch.size"; +241 public static final int DEFAULT_HBASE_REGIONSERVER_MINIBATCH_SIZE = 2; 242 -243 final AtomicBoolean closed = new AtomicBoolean(false); -244 -245 /* Closing can take some time; use the closing flag if there is stuff we don't -246 * want to do while in closing state; e.g. like offer this region up to the -247 * master as a region to close if the carrying regionserver is overloaded. -248 * Once set, it is never cleared. -249 */ -250 final AtomicBoolean closing = new AtomicBoolean(false); -251 -252 /** -253 * The max sequence id of flushed data on this region. There is no edit in memory that is -254 * less that this sequence id. -255 */ -256 private volatile long maxFlushedSeqId = HConstants.NO_SEQNUM; -257 -258 /** -259 * Record the sequence id of last flush operation. Can be in advance of -260 * {@link #maxFlushedSeqId} when flushing a single column family. In this case, -261 * {@link #maxFlushedSeqId} will be older than the oldest edit in memory. -262 */ -263 private volatile long lastFlushOpSeqId = HConstants.NO_SEQNUM; -264 -265 /** -266 * The sequence id of the last replayed open region event from the primary region. This is used -267 * to skip entries before this due to the possibility of replay edits coming out of order from -268 * replication. -269 */ -270 protected volatile long lastReplayedOpenRegionSeqId = -1L; -271 protected volatile long lastReplayedCompactionSeqId = -1L; -272 -273 // -274 // Members -275 // -276 -277 // map from a locked row to the context for that lock including: -278 // - CountDownLatch for threads waiting on that row -279 // - the thread that owns the lock (allow reentrancy) -280 // - reference count of (reentrant) locks held by the thread -281 // - the row itself -282 private final ConcurrentHashMapHashedBytes, RowLockContext lockedRows = -283 new ConcurrentHashMap(); -284 -285 protected final Mapbyte[], HStore stores = -286 new ConcurrentSkipListMap(Bytes.BYTES_RAWCOMPARATOR); +243 public static final String WAL_HSYNC_CONF_KEY = "hbase.wal.hsync"; +244 public static final boolean DEFAULT_WAL_HSYNC = false; +245 +246 final AtomicBoolean closed = new AtomicBoolean(false); +247 +248 /* Closing can take some time; use the closing flag if there is stuff we don't +249 * want to do while in closing state; e.g. like offer this region up to the +250 * master as a region to close if the carrying regionserver is overloaded. +251 * Once set, it is never cleared. +252 */ +253 final AtomicBoolean closing = new AtomicBoolean(false); +254 +255 /** +256 * The max sequence id of flushed data on this region. There is no edit in memory that is +257 * less that this sequence id. +258 */ +259 private volatile long maxFlushedSeqId = HConstants.NO_SEQNUM; +260 +261 /** +262 * Record the sequence id of last flush operation. Can be in advance of +263 * {@link #maxFlushedSeqId} when flushing a single column family. In this case, +264 * {@link #maxFlushedSeqId} will be older than the oldest edit in memory. +265 */ +266 private volatile long
[23/51] [partial] hbase-site git commit: Published site at 00095a2ef9442e3fd86c04876c9d91f2f8b23ad8.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bd675fa3/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRequestFutureImpl.Retry.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRequestFutureImpl.Retry.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRequestFutureImpl.Retry.html index c27b109..4160a88 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRequestFutureImpl.Retry.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRequestFutureImpl.Retry.html @@ -105,7 +105,7 @@ 097try { 098 done = waitUntilDone(startTime * 1000L + asyncProcess.primaryCallTimeoutMicroseconds); 099} catch (InterruptedException ex) { -100 LOG.error("Replica thread was interrupted - no replica calls: " + ex.getMessage()); +100 LOG.error("Replica thread interrupted - no replica calls {}", ex.getMessage()); 101 return; 102} 103 } @@ -149,7 +149,7 @@ 141 if (loc == null) return; 142 HRegionLocation[] locs = loc.getRegionLocations(); 143 if (locs.length == 1) { -144LOG.warn("No replicas found for " + action.getAction()); +144LOG.warn("No replicas found for {}", action.getAction()); 145return; 146 } 147 synchronized (replicaResultLock) { @@ -230,8 +230,8 @@ 222 return; 223} catch (Throwable t) { 224 // This should not happen. Let's log retry anyway. -225 LOG.error("#" + asyncProcess.id + ", Caught throwable while calling. This is unexpected." + -226 " Retrying. Server is " + server + ", tableName=" + tableName, t); +225 LOG.error("id=" + asyncProcess.id + ", caught throwable. Unexpected." + +226 " Retrying. Server=" + server + ", tableName=" + tableName, t); 227 receiveGlobalFailure(multiAction, server, numAttempt, t); 228 return; 229} @@ -247,1036 +247,1035 @@ 239} 240 } catch (Throwable t) { 241// Something really bad happened. We are on the send thread that will now die. -242LOG.error("Internal AsyncProcess #" + asyncProcess.id + " error for " -243+ tableName + " processing for " + server, t); -244throw new RuntimeException(t); -245 } finally { -246 asyncProcess.decTaskCounters(multiAction.getRegions(), server); -247if (callsInProgress != null callable != null res != null) { -248 callsInProgress.remove(callable); -249} -250 } -251} -252 } -253 -254 private final Batch.CallbackCResult callback; -255 private final BatchErrors errors; -256 private final ConnectionImplementation.ServerErrorTracker errorsByServer; -257 private final ExecutorService pool; -258 private final SetCancellableRegionServerCallable callsInProgress; +242LOG.error("id=" + asyncProcess.id + " error for " + tableName + " processing " + server, t); +243throw new RuntimeException(t); +244 } finally { +245 asyncProcess.decTaskCounters(multiAction.getRegions(), server); +246if (callsInProgress != null callable != null res != null) { +247 callsInProgress.remove(callable); +248} +249 } +250} +251 } +252 +253 private final Batch.CallbackCResult callback; +254 private final BatchErrors errors; +255 private final ConnectionImplementation.ServerErrorTracker errorsByServer; +256 private final ExecutorService pool; +257 private final SetCancellableRegionServerCallable callsInProgress; +258 259 -260 -261 private final TableName tableName; -262 private final AtomicLong actionsInProgress = new AtomicLong(-1); -263 /** -264 * The lock controls access to results. It is only held when populating results where -265 * there might be several callers (eventual consistency gets). For other requests, -266 * there's one unique call going on per result index. -267 */ -268 private final Object replicaResultLock = new Object(); -269 /** -270 * Result array. Null if results are not needed. Otherwise, each index corresponds to -271 * the action index in initial actions submitted. For most request types, has null-s for -272 * requests that are not done, and result/exception for those that are done. -273 * For eventual-consistency gets, initially the same applies; at some point, replica calls -274 * might be started, and ReplicaResultState is put at the corresponding indices. The -275 * returning calls check the type to detect when this is the case. After all calls are done, -276 * ReplicaResultState-s are replaced with results for the user. -277 */ -278 private final Object[] results; -279 /** -280 * Indices of replica gets in results. If null, all or no actions are replica-gets. -281 */ -282 private final int[] replicaGetIndices; -283 private final boolean
[23/51] [partial] hbase-site git commit: Published site at 22f4def942f8a3367d0ca6598317e9b9a7d0cfcd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8b1eaec1/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.OperationStatusCode.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.OperationStatusCode.html b/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.OperationStatusCode.html index 26b9cc0..6fa64ff 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.OperationStatusCode.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.OperationStatusCode.html @@ -949,394 +949,411 @@ 941*/ 942 public static final float HBASE_CLUSTER_MINIMUM_MEMORY_THRESHOLD = 0.2f; 943 -944 public static final Pattern CP_HTD_ATTR_KEY_PATTERN = -945 Pattern.compile("^coprocessor\\$([0-9]+)$", Pattern.CASE_INSENSITIVE); -946 -947 /** -948 * pre -949 * Pattern that matches a coprocessor specification. Form is: -950 * {@code coprocessor jar file location '|' class name ['|' priority ['|' arguments]]} -951 * where arguments are {@code KEY '=' VALUE [,...]} -952 * For example: {@code hdfs:///foo.jar|com.foo.FooRegionObserver|1001|arg1=1,arg2=2} -953 * /pre -954 */ -955 public static final Pattern CP_HTD_ATTR_VALUE_PATTERN = -956 Pattern.compile("(^[^\\|]*)\\|([^\\|]+)\\|[\\s]*([\\d]*)[\\s]*(\\|.*)?$"); -957 -958 public static final String CP_HTD_ATTR_VALUE_PARAM_KEY_PATTERN = "[^=,]+"; -959 public static final String CP_HTD_ATTR_VALUE_PARAM_VALUE_PATTERN = "[^,]+"; -960 public static final Pattern CP_HTD_ATTR_VALUE_PARAM_PATTERN = Pattern.compile( -961 "(" + CP_HTD_ATTR_VALUE_PARAM_KEY_PATTERN + ")=(" + -962 CP_HTD_ATTR_VALUE_PARAM_VALUE_PATTERN + "),?"); -963 public static final String CP_HTD_ATTR_INCLUSION_KEY = -964 "hbase.coprocessor.classloader.included.classes"; -965 -966 /** The delay when re-trying a socket operation in a loop (HBASE-4712) */ -967 public static final int SOCKET_RETRY_WAIT_MS = 200; -968 -969 /** Host name of the local machine */ -970 public static final String LOCALHOST = "localhost"; -971 -972 /** -973 * If this parameter is set to true, then hbase will read -974 * data and then verify checksums. Checksum verification -975 * inside hdfs will be switched off. However, if the hbase-checksum -976 * verification fails, then it will switch back to using -977 * hdfs checksums for verifiying data that is being read from storage. -978 * -979 * If this parameter is set to false, then hbase will not -980 * verify any checksums, instead it will depend on checksum verification -981 * being done in the hdfs client. -982 */ -983 public static final String HBASE_CHECKSUM_VERIFICATION = -984 "hbase.regionserver.checksum.verify"; +944 /** +945 * @deprecated It is used internally. As of release 2.0.0, this will be removed in HBase 3.0.0. +946 */ +947 @Deprecated +948 public static final Pattern CP_HTD_ATTR_KEY_PATTERN = +949 Pattern.compile("^coprocessor\\$([0-9]+)$", Pattern.CASE_INSENSITIVE); +950 +951 /** +952 * pre +953 * Pattern that matches a coprocessor specification. Form is: +954 * {@code coprocessor jar file location '|' class name ['|' priority ['|' arguments]]} +955 * where arguments are {@code KEY '=' VALUE [,...]} +956 * For example: {@code hdfs:///foo.jar|com.foo.FooRegionObserver|1001|arg1=1,arg2=2} +957 * /pre +958 * @deprecated It is used internally. As of release 2.0.0, this will be removed in HBase 3.0.0. +959 */ +960 @Deprecated +961 public static final Pattern CP_HTD_ATTR_VALUE_PATTERN = +962 Pattern.compile("(^[^\\|]*)\\|([^\\|]+)\\|[\\s]*([\\d]*)[\\s]*(\\|.*)?$"); +963 /** +964 * @deprecated It is used internally. As of release 2.0.0, this will be removed in HBase 3.0.0. +965 */ +966 @Deprecated +967 public static final String CP_HTD_ATTR_VALUE_PARAM_KEY_PATTERN = "[^=,]+"; +968 /** +969 * @deprecated It is used internally. As of release 2.0.0, this will be removed in HBase 3.0.0. +970 */ +971 @Deprecated +972 public static final String CP_HTD_ATTR_VALUE_PARAM_VALUE_PATTERN = "[^,]+"; +973 /** +974 * @deprecated It is used internally. As of release 2.0.0, this will be removed in HBase 3.0.0. +975 */ +976 @Deprecated +977 public static final Pattern CP_HTD_ATTR_VALUE_PARAM_PATTERN = Pattern.compile( +978 "(" + CP_HTD_ATTR_VALUE_PARAM_KEY_PATTERN + ")=(" + +979 CP_HTD_ATTR_VALUE_PARAM_VALUE_PATTERN + "),?"); +980 public static final String CP_HTD_ATTR_INCLUSION_KEY = +981 "hbase.coprocessor.classloader.included.classes"; +982 +983 /** The delay when re-trying a socket operation in a loop (HBASE-4712) */ +984 public static final int SOCKET_RETRY_WAIT_MS = 200; 985 -986 public static final String LOCALHOST_IP = "127.0.0.1"; -987 -988 public static final String REGION_SERVER_HANDLER_COUNT = "hbase.regionserver.handler.count"; -989 public static final
[23/51] [partial] hbase-site git commit: Published site at 31da4d0bce69b3a47066a5df675756087ce4dc60.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a754d895/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html index 9ae4f41..5950b5e 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/ModifyTableProcedure.html @@ -129,7 +129,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public class ModifyTableProcedure +public class ModifyTableProcedure extends AbstractStateMachineTableProcedureorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableState @@ -400,7 +400,7 @@ extends AbstractStateMachineTableProcedure -acquireLock, checkTableModifiable, getRegionDir, getUser, releaseLock, < a href="../../../../../../org/apache/hadoop/hbase/master/procedure/AbstractStateMachineTableProcedure.html#releaseSyncLatch--">releaseSyncLatch, setUser, toStringClassDetails +acquireLock, checkTableModifiable, getRegionDir, getUser, preflightChecks, releaseLock, releaseSyncLatch, setUser, toStringClassDetails @@ -443,7 +443,7 @@ extends LOG -private static finalorg.slf4j.Logger LOG +private static finalorg.slf4j.Logger LOG @@ -452,7 +452,7 @@ extends unmodifiedTableDescriptor -privateTableDescriptor unmodifiedTableDescriptor +privateTableDescriptor unmodifiedTableDescriptor @@ -461,7 +461,7 @@ extends modifiedTableDescriptor -privateTableDescriptor modifiedTableDescriptor +privateTableDescriptor modifiedTableDescriptor @@ -470,7 +470,7 @@ extends deleteColumnFamilyInModify -privateboolean deleteColumnFamilyInModify +privateboolean deleteColumnFamilyInModify @@ -479,7 +479,7 @@ extends regionInfoList -privatehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo regionInfoList +privatehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo regionInfoList @@ -488,7 +488,7 @@ extends traceEnabled -privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean traceEnabled +privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean traceEnabled @@ -505,7 +505,7 @@ extends ModifyTableProcedure -publicModifyTableProcedure() +publicModifyTableProcedure() @@ -514,8 +514,13 @@ extends ModifyTableProcedure -publicModifyTableProcedure(MasterProcedureEnvenv, -TableDescriptorhtd) +publicModifyTableProcedure(MasterProcedureEnvenv, +TableDescriptorhtd) + throws HBaseIOException + +Throws: +HBaseIOException + @@ -524,9 +529,14 @@ extends ModifyTableProcedure -publicModifyTableProcedure(MasterProcedureEnvenv, +publicModifyTableProcedure(MasterProcedureEnvenv, TableDescriptorhtd, -ProcedurePrepareLatchlatch) +ProcedurePrepareLatchlatch) + throws HBaseIOException + +Throws: +HBaseIOException + @@ -543,7 +553,7 @@ extends initilize -privatevoidinitilize() +privatevoidinitilize() @@ -552,7 +562,7 @@ extends executeFromState -protectedStateMachineProcedure.FlowexecuteFromState(MasterProcedureEnvenv, +protectedStateMachineProcedure.FlowexecuteFromState(MasterProcedureEnvenv, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStatestate) throws https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true; title="class or interface in java.lang">InterruptedException Description copied from class:StateMachineProcedure @@ -575,7 +585,7 @@ extends rollbackState -protectedvoidrollbackState(MasterProcedureEnvenv, +protectedvoidrollbackState(MasterProcedureEnvenv, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStatestate) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Description copied from class:StateMachineProcedure @@ -595,7 +605,7 @@ extends isRollbackSupported -protectedbooleanisRollbackSupported(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ModifyTableStatestate)
[23/51] [partial] hbase-site git commit: Published site at 6b77786dfc46d25ac5bb5f1c8a4a9eb47b52a604.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html index 8225386..5955f86 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html @@ -495,7 +495,7 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. private static HRegionLocation -MetaTableAccessor.getRegionLocation(Resultr, +AsyncMetaTableAccessor.getRegionLocation(Resultr, RegionInforegionInfo, intreplicaId) Returns the HRegionLocation parsed from the given meta row Result @@ -504,7 +504,7 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. private static HRegionLocation -AsyncMetaTableAccessor.getRegionLocation(Resultr, +MetaTableAccessor.getRegionLocation(Resultr, RegionInforegionInfo, intreplicaId) Returns the HRegionLocation parsed from the given meta row Result @@ -941,9 +941,7 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -AsyncAdmin.getRegions(ServerNameserverName) -Get all the online regions on a region server. - +AsyncHBaseAdmin.getRegions(ServerNameserverName) https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo @@ -952,22 +950,22 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -RawAsyncHBaseAdmin.getRegions(ServerNameserverName) - - https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo HBaseAdmin.getRegions(ServerNamesn) + +https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo +AsyncAdmin.getRegions(ServerNameserverName) +Get all the online regions on a region server. + + https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -AsyncHBaseAdmin.getRegions(ServerNameserverName) +RawAsyncHBaseAdmin.getRegions(ServerNameserverName) https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -AsyncAdmin.getRegions(TableNametableName) -Get the regions of a given table. - +AsyncHBaseAdmin.getRegions(TableNametableName) https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo @@ -976,16 +974,18 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -RawAsyncHBaseAdmin.getRegions(TableNametableName) - - https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo HBaseAdmin.getRegions(TableNametableName) + +https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo
[23/51] [partial] hbase-site git commit: Published site at 1384da71375427b522b09f06862bb5d629cef52f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d347bde8/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html index 253a995..f484fd7 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Connection.html @@ -798,23 +798,23 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. private Connection -RestoreTablesClient.conn +BackupAdminImpl.conn -protected Connection -TableBackupClient.conn - - (package private) Connection BackupCommands.Command.conn + +private Connection +RestoreTablesClient.conn + protected Connection -BackupManager.conn +TableBackupClient.conn -private Connection -BackupAdminImpl.conn +protected Connection +BackupManager.conn private Connection @@ -1128,13 +1128,13 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -(package private) Connection -ConnectionImplementation.MasterServiceState.connection - - private Connection RegionServerCallable.connection + +(package private) Connection +ConnectionImplementation.MasterServiceState.connection + @@ -1179,20 +1179,20 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -Connection -Admin.getConnection() - - (package private) Connection RegionAdminServiceCallable.getConnection() - + protected Connection HTable.getConnection() INTERNAL Used by unit tests and tools to do low-level manipulations. + +Connection +Admin.getConnection() + Connection HBaseAdmin.getConnection() @@ -1506,11 +1506,11 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. private Connection -TableInputFormatBase.connection +HRegionPartitioner.connection private Connection -HRegionPartitioner.connection +TableInputFormatBase.connection @@ -1543,22 +1543,22 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -private Connection -TableOutputFormat.TableRecordWriter.connection - - (package private) Connection MultiTableOutputFormat.MultiTableRecordWriter.connection + +private Connection +HRegionPartitioner.connection + private Connection -TableInputFormatBase.connection -The underlying Connection of the table. - +TableOutputFormat.TableRecordWriter.connection private Connection -HRegionPartitioner.connection +TableInputFormatBase.connection +The underlying Connection of the table. + (package private) Connection @@ -1643,15 +1643,15 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. private Connection -CatalogJanitor.connection +RegionPlacementMaintainer.connection private Connection -SnapshotOfRegionAssignmentFromMeta.connection +CatalogJanitor.connection private Connection -RegionPlacementMaintainer.connection +SnapshotOfRegionAssignmentFromMeta.connection @@ -1787,16 +1787,16 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -private Connection -TableQuotaSnapshotStore.conn +(package private) Connection +FileArchiverNotifierFactoryImpl.CacheKey.conn private Connection -SpaceQuotaRefresherChore.conn +QuotaObserverChore.conn private Connection -NamespaceQuotaSnapshotStore.conn +QuotaObserverChore.TablesWithQuotas.conn private Connection @@ -1804,11 +1804,11 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. private Connection -QuotaObserverChore.conn +NamespaceQuotaSnapshotStore.conn private Connection -QuotaObserverChore.TablesWithQuotas.conn +TableQuotaSnapshotStore.conn private Connection @@ -1816,6 +1816,14 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. private Connection +FileArchiverNotifierImpl.conn + + +private Connection +SpaceQuotaRefresherChore.conn + + +private Connection QuotaRetriever.connection Connection to use. @@ -1972,6 +1980,24 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListGetgets) +FileArchiverNotifier +FileArchiverNotifierFactoryImpl.get(Connectionconn, + org.apache.hadoop.conf.Configurationconf, + org.apache.hadoop.fs.FileSystemfs, + TableNametn) +Returns the FileArchiverNotifier instance for the given TableName. + + + +FileArchiverNotifier +FileArchiverNotifierFactory.get(Connectionconn, + org.apache.hadoop.conf.Configurationconf, + org.apache.hadoop.fs.FileSystemfs, + TableNametn) +Creates or obtains a FileArchiverNotifier instance for the given args. + + + static SpaceQuotaSnapshot QuotaTableUtil.getCurrentSnapshot(Connectionconn,
[23/51] [partial] hbase-site git commit: Published site at b7b86839250bf9b295ebc1948826f43a88736d6c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6b94a2f2/devapidocs/org/apache/hadoop/hbase/util/FSUtils.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/util/FSUtils.html b/devapidocs/org/apache/hadoop/hbase/util/FSUtils.html index 42307d2..2ade137 100644 --- a/devapidocs/org/apache/hadoop/hbase/util/FSUtils.html +++ b/devapidocs/org/apache/hadoop/hbase/util/FSUtils.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9,"i7":9,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":9,"i26":9,"i27":9,"i28":9,"i29":9,"i30":9,"i31":9,"i32":9,"i33":9,"i34":9,"i35":9,"i36":9,"i37":9,"i38":9,"i39":9,"i40":6,"i41":9,"i42":9,"i43":9,"i44":9,"i45":9,"i46":9,"i47":9,"i48":9}; +var methods = {"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9,"i7":9,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":9,"i26":9,"i27":9,"i28":9,"i29":9,"i30":9,"i31":9,"i32":9,"i33":9,"i34":9,"i35":9,"i36":9,"i37":9,"i38":9,"i39":9,"i40":9,"i41":6,"i42":9,"i43":9,"i44":9,"i45":9,"i46":9,"i47":9,"i48":9,"i49":9}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -429,6 +429,11 @@ extends +static org.apache.hadoop.fs.Path +getRegionDir(org.apache.hadoop.fs.PathtableDir, +RegionInforegion) + + static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.fs.Path getRegionDirs(org.apache.hadoop.fs.FileSystemfs, org.apache.hadoop.fs.PathtableDir) @@ -436,7 +441,7 @@ extends - + private static void getRegionLocalityMappingFromFS(org.apache.hadoop.conf.Configurationconf, https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringdesiredTable, @@ -449,17 +454,17 @@ extends - + static int getRegionReferenceFileCount(org.apache.hadoop.fs.FileSystemfs, org.apache.hadoop.fs.Pathp) - + static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.fs.Path getTableDirs(org.apache.hadoop.fs.FileSystemfs, org.apache.hadoop.fs.Pathrootdir) - + static https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer getTableFragmentation(org.apache.hadoop.fs.FileSystemfs, org.apache.hadoop.fs.PathhbaseRootDir) @@ -467,14 +472,14 @@ extends - + static https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer getTableFragmentation(HMastermaster) Runs through the HBase rootdir and checks how many stores for each table have more than one file in them. - + static https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,org.apache.hadoop.fs.Path getTableStoreFilePathMap(org.apache.hadoop.fs.FileSystemfs, org.apache.hadoop.fs.PathhbaseRootDir) @@ -482,7 +487,7 @@ extends - + static https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,org.apache.hadoop.fs.Path getTableStoreFilePathMap(org.apache.hadoop.fs.FileSystemfs, org.apache.hadoop.fs.PathhbaseRootDir, @@ -493,7 +498,7 @@ extends - + static https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,org.apache.hadoop.fs.Path
[23/51] [partial] hbase-site git commit: Published site at 1d25b60831b8cc8f7ad5fd366f1867de5c20d2f3.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb05e3e3/apidocs/org/apache/hadoop/hbase/client/AdvancedScanResultConsumer.ScanController.html -- diff --git a/apidocs/org/apache/hadoop/hbase/client/AdvancedScanResultConsumer.ScanController.html b/apidocs/org/apache/hadoop/hbase/client/AdvancedScanResultConsumer.ScanController.html index e12d889..afdcd68 100644 --- a/apidocs/org/apache/hadoop/hbase/client/AdvancedScanResultConsumer.ScanController.html +++ b/apidocs/org/apache/hadoop/hbase/client/AdvancedScanResultConsumer.ScanController.html @@ -134,7 +134,7 @@ public static interface Method and Description -http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">OptionalCursor +https://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">OptionalCursor cursor() Get the scan cursor if available. @@ -202,7 +202,7 @@ public static interface cursor -http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">OptionalCursorcursor() +https://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">OptionalCursorcursor() Get the scan cursor if available. Returns: http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb05e3e3/apidocs/org/apache/hadoop/hbase/client/Append.html -- diff --git a/apidocs/org/apache/hadoop/hbase/client/Append.html b/apidocs/org/apache/hadoop/hbase/client/Append.html index e41805f..b655827 100644 --- a/apidocs/org/apache/hadoop/hbase/client/Append.html +++ b/apidocs/org/apache/hadoop/hbase/client/Append.html @@ -97,7 +97,7 @@ var activeTableTab = "activeTableTab"; -http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object org.apache.hadoop.hbase.client.Operation @@ -124,7 +124,7 @@ var activeTableTab = "activeTableTab"; All Implemented Interfaces: -http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableRow, CellScannable, Attributes, Row, org.apache.hadoop.hbase.io.HeapSize +https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableRow, CellScannable, Attributes, Row, org.apache.hadoop.hbase.io.HeapSize @@ -206,7 +206,7 @@ extends Append(byte[]row, longts, - http://docs.oracle.com/javase/8/docs/api/java/util/NavigableMap.html?is-external=true; title="class or interface in java.util">NavigableMapbyte[],http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListCellfamilyMap) + https://docs.oracle.com/javase/8/docs/api/java/util/NavigableMap.html?is-external=true; title="class or interface in java.util">NavigableMapbyte[],https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListCellfamilyMap) Construct the Append with user defined data. @@ -268,16 +268,16 @@ extends Append -setACL(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,org.apache.hadoop.hbase.security.access.Permissionperms) +setACL(https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,org.apache.hadoop.hbase.security.access.Permissionperms) Append -setACL(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringuser, +setACL(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringuser, org.apache.hadoop.hbase.security.access.Permissionperms) Append -setAttribute(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname, +setAttribute(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname, byte[]value) Sets an attribute. @@ -290,7 +290,7 @@ extends Append
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ec8bf761/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerThread.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerThread.html b/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerThread.html index 6cba39a..3a47680 100644 --- a/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerThread.html +++ b/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerThread.html @@ -103,7 +103,7 @@ var activeTableTab = "activeTableTab"; http://docs.oracle.com/javase/8/docs/api/java/lang/Thread.html?is-external=true; title="class or interface in java.lang">java.lang.Thread -org.apache.hadoop.hbase.procedure2.ProcedureExecutor.StoppableThread +org.apache.hadoop.hbase.procedure2.StoppableThread org.apache.hadoop.hbase.procedure2.ProcedureExecutor.WorkerThread @@ -122,13 +122,17 @@ var activeTableTab = "activeTableTab"; http://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true; title="class or interface in java.lang">Runnable +Direct Known Subclasses: +ProcedureExecutor.KeepAliveWorkerThread + + Enclosing class: ProcedureExecutorTEnvironment -private final class ProcedureExecutor.WorkerThread -extends ProcedureExecutor.StoppableThread +private class ProcedureExecutor.WorkerThread +extends StoppableThread @@ -163,7 +167,7 @@ extends Field and Description -private Procedure +private Procedure? activeProcedure @@ -189,10 +193,17 @@ extends Constructors -Constructor and Description +Modifier +Constructor and Description -WorkerThread(http://docs.oracle.com/javase/8/docs/api/java/lang/ThreadGroup.html?is-external=true; title="class or interface in java.lang">ThreadGroupgroup) + +WorkerThread(http://docs.oracle.com/javase/8/docs/api/java/lang/ThreadGroup.html?is-external=true; title="class or interface in java.lang">ThreadGroupgroup) + + +protected +WorkerThread(http://docs.oracle.com/javase/8/docs/api/java/lang/ThreadGroup.html?is-external=true; title="class or interface in java.lang">ThreadGroupgroup, +http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringprefix) @@ -214,7 +225,7 @@ extends getCurrentRunTime() -private boolean +protected boolean keepAlive(longlastUpdate) @@ -231,11 +242,11 @@ extends - + -Methods inherited from classorg.apache.hadoop.hbase.procedure2.ProcedureExecutor.StoppableThread -awaitTermination +Methods inherited from classorg.apache.hadoop.hbase.procedure2.StoppableThread +awaitTermination @@ -271,7 +282,7 @@ extends executionStartTime -private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicLong.html?is-external=true; title="class or interface in java.util.concurrent.atomic">AtomicLong executionStartTime +private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicLong.html?is-external=true; title="class or interface in java.util.concurrent.atomic">AtomicLong executionStartTime @@ -280,7 +291,7 @@ extends activeProcedure -privateProcedure activeProcedure +private volatileProcedure? activeProcedure @@ -294,10 +305,20 @@ extends + + +WorkerThread +publicWorkerThread(http://docs.oracle.com/javase/8/docs/api/java/lang/ThreadGroup.html?is-external=true; title="class or interface in java.lang">ThreadGroupgroup) + + + + + WorkerThread -publicWorkerThread(http://docs.oracle.com/javase/8/docs/api/java/lang/ThreadGroup.html?is-external=true; title="class or interface in java.lang">ThreadGroupgroup) +protectedWorkerThread(http://docs.oracle.com/javase/8/docs/api/java/lang/ThreadGroup.html?is-external=true; title="class or interface in java.lang">ThreadGroupgroup, + http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringprefix) @@ -314,10 +335,10 @@ extends sendStopSignal -publicvoidsendStopSignal() +publicvoidsendStopSignal() Specified by: -sendStopSignalin classProcedureExecutor.StoppableThread +sendStopSignalin classStoppableThread @@ -327,7 +348,7 @@ extends run -publicvoidrun() +publicvoidrun() Specified by: http://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true#run--; title="class or interface in java.lang">runin interfacehttp://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true; title="class or interface in java.lang">Runnable @@ -342,7 +363,7 @@ extends toString -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringtoString() +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/991224b9/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html index 70c0a84..f458c4c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html @@ -263,541 +263,547 @@ 255throw new UnsupportedOperationException(this + " unhandled state=" + state); 256 } 257} catch (IOException e) { -258 LOG.warn("Error trying to merge regions " + RegionInfo.getShortNameToLog(regionsToMerge) + -259" in the table " + getTableName() + " (in state=" + state + ")", e); -260 -261 setFailure("master-merge-regions", e); -262} -263return Flow.HAS_MORE_STATE; -264 } -265 -266 /** -267 * To rollback {@link MergeTableRegionsProcedure}, two AssignProcedures are asynchronously -268 * submitted for each region to be merged (rollback doesn't wait on the completion of the -269 * AssignProcedures) . This can be improved by changing rollback() to support sub-procedures. -270 * See HBASE-19851 for details. -271 */ -272 @Override -273 protected void rollbackState( -274 final MasterProcedureEnv env, -275 final MergeTableRegionsState state) throws IOException, InterruptedException { -276if (isTraceEnabled()) { -277 LOG.trace(this + " rollback state=" + state); -278} -279 -280try { -281 switch (state) { -282 case MERGE_TABLE_REGIONS_POST_OPERATION: -283 case MERGE_TABLE_REGIONS_OPEN_MERGED_REGION: -284 case MERGE_TABLE_REGIONS_POST_MERGE_COMMIT_OPERATION: -285 case MERGE_TABLE_REGIONS_UPDATE_META: -286String msg = this + " We are in the " + state + " state." -287+ " It is complicated to rollback the merge operation that region server is working on." -288+ " Rollback is not supported and we should let the merge operation to complete"; -289LOG.warn(msg); -290// PONR -291throw new UnsupportedOperationException(this + " unhandled state=" + state); -292 case MERGE_TABLE_REGIONS_PRE_MERGE_COMMIT_OPERATION: -293break; -294 case MERGE_TABLE_REGIONS_CREATE_MERGED_REGION: -295cleanupMergedRegion(env); -296break; -297 case MERGE_TABLE_REGIONS_CLOSE_REGIONS: -298 rollbackCloseRegionsForMerge(env); +258 String msg = "Error trying to merge regions " + +259 RegionInfo.getShortNameToLog(regionsToMerge) + " in the table " + getTableName() + +260 " (in state=" + state + ")"; +261 if (!isRollbackSupported(state)) { +262// We reach a state that cannot be rolled back. We just need to keep retry. +263LOG.warn(msg, e); +264 } else { +265LOG.error(msg, e); +266 setFailure("master-merge-regions", e); +267 } +268} +269return Flow.HAS_MORE_STATE; +270 } +271 +272 /** +273 * To rollback {@link MergeTableRegionsProcedure}, two AssignProcedures are asynchronously +274 * submitted for each region to be merged (rollback doesn't wait on the completion of the +275 * AssignProcedures) . This can be improved by changing rollback() to support sub-procedures. +276 * See HBASE-19851 for details. +277 */ +278 @Override +279 protected void rollbackState( +280 final MasterProcedureEnv env, +281 final MergeTableRegionsState state) throws IOException, InterruptedException { +282if (isTraceEnabled()) { +283 LOG.trace(this + " rollback state=" + state); +284} +285 +286try { +287 switch (state) { +288 case MERGE_TABLE_REGIONS_POST_OPERATION: +289 case MERGE_TABLE_REGIONS_OPEN_MERGED_REGION: +290 case MERGE_TABLE_REGIONS_POST_MERGE_COMMIT_OPERATION: +291 case MERGE_TABLE_REGIONS_UPDATE_META: +292String msg = this + " We are in the " + state + " state." +293+ " It is complicated to rollback the merge operation that region server is working on." +294+ " Rollback is not supported and we should let the merge operation to complete"; +295LOG.warn(msg); +296// PONR +297throw new UnsupportedOperationException(this + " unhandled state=" + state); +298 case MERGE_TABLE_REGIONS_PRE_MERGE_COMMIT_OPERATION: 299break; -300 case MERGE_TABLE_REGIONS_PRE_MERGE_OPERATION: -301postRollBackMergeRegions(env); +300 case MERGE_TABLE_REGIONS_CREATE_MERGED_REGION: +301cleanupMergedRegion(env); 302break; -303 case MERGE_TABLE_REGIONS_PREPARE: -304break; -305 default: -306throw new
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html index 6d8c565..7edb3ff 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html @@ -2726,863 +2726,873 @@ 2718} 2719 } 2720 -2721 void checkInitialized() -2722 throws PleaseHoldException, ServerNotRunningYetException, MasterNotRunningException { -2723checkServiceStarted(); -2724if (!isInitialized()) throw new PleaseHoldException("Master is initializing"); -2725if (isStopped()) throw new MasterNotRunningException(); -2726 } -2727 -2728 /** -2729 * Report whether this master is currently the active master or not. -2730 * If not active master, we are parked on ZK waiting to become active. -2731 * -2732 * This method is used for testing. -2733 * -2734 * @return true if active master, false if not. -2735 */ -2736 @Override -2737 public boolean isActiveMaster() { -2738return activeMaster; -2739 } -2740 -2741 /** -2742 * Report whether this master has completed with its initialization and is -2743 * ready. If ready, the master is also the active master. A standby master -2744 * is never ready. -2745 * -2746 * This method is used for testing. -2747 * -2748 * @return true if master is ready to go, false if not. -2749 */ -2750 @Override -2751 public boolean isInitialized() { -2752return initialized.isReady(); -2753 } -2754 -2755 /** -2756 * Report whether this master is in maintenance mode. +2721 public static class MasterStoppedException extends DoNotRetryIOException { +2722MasterStoppedException() { +2723 super(); +2724} +2725 } +2726 +2727 void checkInitialized() throws PleaseHoldException, ServerNotRunningYetException, +2728 MasterNotRunningException, MasterStoppedException { +2729checkServiceStarted(); +2730if (!isInitialized()) { +2731 throw new PleaseHoldException("Master is initializing"); +2732} +2733if (isStopped()) { +2734 throw new MasterStoppedException(); +2735} +2736 } +2737 +2738 /** +2739 * Report whether this master is currently the active master or not. +2740 * If not active master, we are parked on ZK waiting to become active. +2741 * +2742 * This method is used for testing. +2743 * +2744 * @return true if active master, false if not. +2745 */ +2746 @Override +2747 public boolean isActiveMaster() { +2748return activeMaster; +2749 } +2750 +2751 /** +2752 * Report whether this master has completed with its initialization and is +2753 * ready. If ready, the master is also the active master. A standby master +2754 * is never ready. +2755 * +2756 * This method is used for testing. 2757 * -2758 * @return true if master is in maintenanceMode +2758 * @return true if master is ready to go, false if not. 2759 */ 2760 @Override -2761 public boolean isInMaintenanceMode() { -2762return maintenanceModeTracker.isInMaintenanceMode(); +2761 public boolean isInitialized() { +2762return initialized.isReady(); 2763 } 2764 -2765 @VisibleForTesting -2766 public void setInitialized(boolean isInitialized) { -2767 procedureExecutor.getEnvironment().setEventReady(initialized, isInitialized); -2768 } -2769 +2765 /** +2766 * Report whether this master is in maintenance mode. +2767 * +2768 * @return true if master is in maintenanceMode +2769 */ 2770 @Override -2771 public ProcedureEvent? getInitializedEvent() { -2772return initialized; +2771 public boolean isInMaintenanceMode() { +2772return maintenanceModeTracker.isInMaintenanceMode(); 2773 } 2774 -2775 /** -2776 * ServerCrashProcessingEnabled is set false before completing assignMeta to prevent processing -2777 * of crashed servers. -2778 * @return true if assignMeta has completed; -2779 */ +2775 @VisibleForTesting +2776 public void setInitialized(boolean isInitialized) { +2777 procedureExecutor.getEnvironment().setEventReady(initialized, isInitialized); +2778 } +2779 2780 @Override -2781 public boolean isServerCrashProcessingEnabled() { -2782return serverCrashProcessingEnabled.isReady(); +2781 public ProcedureEvent? getInitializedEvent() { +2782return initialized; 2783 } 2784 -2785 @VisibleForTesting -2786 public void setServerCrashProcessingEnabled(final boolean b) { -2787 procedureExecutor.getEnvironment().setEventReady(serverCrashProcessingEnabled, b); -2788 } -2789 -2790 public ProcedureEvent? getServerCrashProcessingEnabledEvent() { -2791return
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/94208cfe/devapidocs/org/apache/hadoop/hbase/client/class-use/SnapshotDescription.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/SnapshotDescription.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/SnapshotDescription.html index 4584cda..fb9bdb3 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/SnapshotDescription.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/SnapshotDescription.html @@ -137,9 +137,7 @@ http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListSnapshotDescription -AsyncAdmin.listSnapshots() -List completed snapshots. - +AsyncHBaseAdmin.listSnapshots() http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListSnapshotDescription @@ -148,22 +146,22 @@ -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListSnapshotDescription -RawAsyncHBaseAdmin.listSnapshots() - - http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListSnapshotDescription HBaseAdmin.listSnapshots() + +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListSnapshotDescription +AsyncAdmin.listSnapshots() +List completed snapshots. + + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListSnapshotDescription -AsyncHBaseAdmin.listSnapshots() +RawAsyncHBaseAdmin.listSnapshots() http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListSnapshotDescription -AsyncAdmin.listSnapshots(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Patternpattern) -List all the completed snapshots matching the given pattern. - +AsyncHBaseAdmin.listSnapshots(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Patternpattern) http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListSnapshotDescription @@ -172,16 +170,18 @@ -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListSnapshotDescription -RawAsyncHBaseAdmin.listSnapshots(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Patternpattern) - - http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListSnapshotDescription HBaseAdmin.listSnapshots(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Patternpattern) + +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListSnapshotDescription +AsyncAdmin.listSnapshots(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Patternpattern) +List all the completed snapshots matching the given pattern. + + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cd17dc5/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html index 232a8b4..b0d9cb7 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RegionInfo.html @@ -495,7 +495,7 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. private static HRegionLocation -AsyncMetaTableAccessor.getRegionLocation(Resultr, +MetaTableAccessor.getRegionLocation(Resultr, RegionInforegionInfo, intreplicaId) Returns the HRegionLocation parsed from the given meta row Result @@ -504,7 +504,7 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. private static HRegionLocation -MetaTableAccessor.getRegionLocation(Resultr, +AsyncMetaTableAccessor.getRegionLocation(Resultr, RegionInforegionInfo, intreplicaId) Returns the HRegionLocation parsed from the given meta row Result @@ -944,7 +944,9 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -AsyncHBaseAdmin.getRegions(ServerNameserverName) +AsyncAdmin.getRegions(ServerNameserverName) +Get all the online regions on a region server. + http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo @@ -953,22 +955,22 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -HBaseAdmin.getRegions(ServerNamesn) +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo +RawAsyncHBaseAdmin.getRegions(ServerNameserverName) -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -AsyncAdmin.getRegions(ServerNameserverName) -Get all the online regions on a region server. - +http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo +HBaseAdmin.getRegions(ServerNamesn) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -RawAsyncHBaseAdmin.getRegions(ServerNameserverName) +AsyncHBaseAdmin.getRegions(ServerNameserverName) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -AsyncHBaseAdmin.getRegions(TableNametableName) +AsyncAdmin.getRegions(TableNametableName) +Get the regions of a given table. + http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo @@ -977,18 +979,16 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo -HBaseAdmin.getRegions(TableNametableName) +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo +RawAsyncHBaseAdmin.getRegions(TableNametableName) -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html index 35f0e35..e3d9f70 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html @@ -208,9 +208,9 @@ service. -ResultScanner -HTable.getScanner(byte[]family) -The underlying HTable must not be closed. +default ResultScanner +AsyncTable.getScanner(byte[]family) +Gets a scanner on the current table for the given family. @@ -220,16 +220,16 @@ service. -default ResultScanner -AsyncTable.getScanner(byte[]family) -Gets a scanner on the current table for the given family. +ResultScanner +HTable.getScanner(byte[]family) +The underlying HTable must not be closed. -ResultScanner -HTable.getScanner(byte[]family, +default ResultScanner +AsyncTable.getScanner(byte[]family, byte[]qualifier) -The underlying HTable must not be closed. +Gets a scanner on the current table for the given family and qualifier. @@ -240,37 +240,37 @@ service. -default ResultScanner -AsyncTable.getScanner(byte[]family, +ResultScanner +HTable.getScanner(byte[]family, byte[]qualifier) -Gets a scanner on the current table for the given family and qualifier. +The underlying HTable must not be closed. ResultScanner -RawAsyncTableImpl.getScanner(Scanscan) - - -ResultScanner -HTable.getScanner(Scanscan) -The underlying HTable must not be closed. +AsyncTable.getScanner(Scanscan) +Returns a scanner on the current table as specified by the Scan object. - + ResultScanner Table.getScanner(Scanscan) Returns a scanner on the current table as specified by the Scan object. - + ResultScanner AsyncTableImpl.getScanner(Scanscan) + +ResultScanner +RawAsyncTableImpl.getScanner(Scanscan) + ResultScanner -AsyncTable.getScanner(Scanscan) -Returns a scanner on the current table as specified by the Scan object. +HTable.getScanner(Scanscan) +The underlying HTable must not be closed. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html index d730879..b1d1cef 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html @@ -106,11 +106,11 @@ RetriesExhaustedWithDetailsException -AsyncRequestFutureImpl.getErrors() +AsyncRequestFuture.getErrors() RetriesExhaustedWithDetailsException -AsyncRequestFuture.getErrors() +AsyncRequestFutureImpl.getErrors() (package private) RetriesExhaustedWithDetailsException http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html index 9642faa..0a290e1 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html @@ -234,36 +234,28 @@ -T -RpcRetryingCallerImpl.callWithoutRetries(RetryingCallableTcallable, - intcallTimeout) - - T RpcRetryingCaller.callWithoutRetries(RetryingCallableTcallable, intcallTimeout) Call the server once only. - + T -RpcRetryingCallerImpl.callWithRetries(RetryingCallableTcallable, - intcallTimeout) +RpcRetryingCallerImpl.callWithoutRetries(RetryingCallableTcallable, + intcallTimeout) - + T RpcRetryingCaller.callWithRetries(RetryingCallableTcallable, intcallTimeout) Retries if invocation fails. - -RetryingCallerInterceptorContext -NoOpRetryingInterceptorContext.prepare(RetryingCallable?callable) - -FastFailInterceptorContext -FastFailInterceptorContext.prepare(RetryingCallable?callable) +T +RpcRetryingCallerImpl.callWithRetries(RetryingCallableTcallable, + intcallTimeout) abstract RetryingCallerInterceptorContext @@ -275,13 +267,11 @@ RetryingCallerInterceptorContext -NoOpRetryingInterceptorContext.prepare(RetryingCallable?callable, - inttries)
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/client/class-use/Get.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Get.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/Get.html index 7676dbd..dd70c74 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Get.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Get.html @@ -399,10 +399,8 @@ service. -default http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean -AsyncTable.exists(Getget) -Test for the existence of columns in the table, as specified by the Get. - +boolean +HTable.exists(Getget) boolean @@ -411,32 +409,34 @@ service. -boolean -HTable.exists(Getget) +default http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean +AsyncTable.exists(Getget) +Test for the existence of columns in the table, as specified by the Get. + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureResult -AsyncTable.get(Getget) -Extracts certain cells from a given row. - +RawAsyncTableImpl.get(Getget) Result +HTable.get(Getget) + + +Result Table.get(Getget) Extracts certain cells from a given row. - -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureResult -AsyncTableImpl.get(Getget) - http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureResult -RawAsyncTableImpl.get(Getget) +AsyncTableImpl.get(Getget) -Result -HTable.get(Getget) +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureResult +AsyncTable.get(Getget) +Extracts certain cells from a given row. + private Result @@ -457,10 +457,8 @@ service. -default http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean -AsyncTable.exists(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListGetgets) -Test for the existence of columns in the table, as specified by the Gets. - +boolean[] +HTable.exists(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListGetgets) boolean[] @@ -469,16 +467,12 @@ service. -boolean[] -HTable.exists(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListGetgets) - - -default http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean -AsyncTable.existsAll(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListGetgets) -A simple version for batch exists. +default http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean +AsyncTable.exists(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListGetgets) +Test for the existence of columns in the table, as specified by the Gets. -
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html index 35f0e35..e3d9f70 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/ResultScanner.html @@ -208,9 +208,9 @@ service. -ResultScanner -HTable.getScanner(byte[]family) -The underlying HTable must not be closed. +default ResultScanner +AsyncTable.getScanner(byte[]family) +Gets a scanner on the current table for the given family. @@ -220,16 +220,16 @@ service. -default ResultScanner -AsyncTable.getScanner(byte[]family) -Gets a scanner on the current table for the given family. +ResultScanner +HTable.getScanner(byte[]family) +The underlying HTable must not be closed. -ResultScanner -HTable.getScanner(byte[]family, +default ResultScanner +AsyncTable.getScanner(byte[]family, byte[]qualifier) -The underlying HTable must not be closed. +Gets a scanner on the current table for the given family and qualifier. @@ -240,37 +240,37 @@ service. -default ResultScanner -AsyncTable.getScanner(byte[]family, +ResultScanner +HTable.getScanner(byte[]family, byte[]qualifier) -Gets a scanner on the current table for the given family and qualifier. +The underlying HTable must not be closed. ResultScanner -RawAsyncTableImpl.getScanner(Scanscan) - - -ResultScanner -HTable.getScanner(Scanscan) -The underlying HTable must not be closed. +AsyncTable.getScanner(Scanscan) +Returns a scanner on the current table as specified by the Scan object. - + ResultScanner Table.getScanner(Scanscan) Returns a scanner on the current table as specified by the Scan object. - + ResultScanner AsyncTableImpl.getScanner(Scanscan) + +ResultScanner +RawAsyncTableImpl.getScanner(Scanscan) + ResultScanner -AsyncTable.getScanner(Scanscan) -Returns a scanner on the current table as specified by the Scan object. +HTable.getScanner(Scanscan) +The underlying HTable must not be closed. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html index d730879..b1d1cef 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetriesExhaustedWithDetailsException.html @@ -106,11 +106,11 @@ RetriesExhaustedWithDetailsException -AsyncRequestFutureImpl.getErrors() +AsyncRequestFuture.getErrors() RetriesExhaustedWithDetailsException -AsyncRequestFuture.getErrors() +AsyncRequestFutureImpl.getErrors() (package private) RetriesExhaustedWithDetailsException http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html index 9642faa..0a290e1 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallable.html @@ -234,36 +234,28 @@ -T -RpcRetryingCallerImpl.callWithoutRetries(RetryingCallableTcallable, - intcallTimeout) - - T RpcRetryingCaller.callWithoutRetries(RetryingCallableTcallable, intcallTimeout) Call the server once only. - + T -RpcRetryingCallerImpl.callWithRetries(RetryingCallableTcallable, - intcallTimeout) +RpcRetryingCallerImpl.callWithoutRetries(RetryingCallableTcallable, + intcallTimeout) - + T RpcRetryingCaller.callWithRetries(RetryingCallableTcallable, intcallTimeout) Retries if invocation fails. - -RetryingCallerInterceptorContext -NoOpRetryingInterceptorContext.prepare(RetryingCallable?callable) - -FastFailInterceptorContext -FastFailInterceptorContext.prepare(RetryingCallable?callable) +T +RpcRetryingCallerImpl.callWithRetries(RetryingCallableTcallable, + intcallTimeout) abstract RetryingCallerInterceptorContext @@ -275,13 +267,11 @@ RetryingCallerInterceptorContext -NoOpRetryingInterceptorContext.prepare(RetryingCallable?callable, - inttries)
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0ab8335e/devapidocs/org/apache/hadoop/hbase/client/class-use/RpcRetryingCallerFactory.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/RpcRetryingCallerFactory.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/RpcRetryingCallerFactory.html index d481372..5e1590b 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RpcRetryingCallerFactory.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RpcRetryingCallerFactory.html @@ -126,15 +126,15 @@ private RpcRetryingCallerFactory -RegionCoprocessorRpcChannel.rpcCallerFactory +ConnectionImplementation.rpcCallerFactory private RpcRetryingCallerFactory -ConnectionImplementation.rpcCallerFactory +HTable.rpcCallerFactory private RpcRetryingCallerFactory -HTable.rpcCallerFactory +RegionCoprocessorRpcChannel.rpcCallerFactory private RpcRetryingCallerFactory @@ -155,21 +155,21 @@ RpcRetryingCallerFactory -ConnectionImplementation.getNewRpcRetryingCallerFactory(org.apache.hadoop.conf.Configurationconf) - - -RpcRetryingCallerFactory ClusterConnection.getNewRpcRetryingCallerFactory(org.apache.hadoop.conf.Configurationconf) Returns a new RpcRetryingCallerFactory from the given Configuration. + +RpcRetryingCallerFactory +ConnectionImplementation.getNewRpcRetryingCallerFactory(org.apache.hadoop.conf.Configurationconf) + RpcRetryingCallerFactory -ConnectionImplementation.getRpcRetryingCallerFactory() +ClusterConnection.getRpcRetryingCallerFactory() RpcRetryingCallerFactory -ClusterConnection.getRpcRetryingCallerFactory() +ConnectionImplementation.getRpcRetryingCallerFactory() static RpcRetryingCallerFactory http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0ab8335e/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html index 6384833..018438c 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Scan.html @@ -283,27 +283,27 @@ service. private Scan -ScannerCallableWithReplicas.scan +AsyncScanSingleRegionRpcRetryingCaller.scan protected Scan -ClientScanner.scan +ScannerCallable.scan private Scan -AsyncClientScanner.scan +ScannerCallableWithReplicas.scan -private Scan -AsyncRpcRetryingCallerFactory.ScanSingleRegionCallerBuilder.scan +protected Scan +ClientScanner.scan private Scan -AsyncScanSingleRegionRpcRetryingCaller.scan +AsyncClientScanner.scan -protected Scan -ScannerCallable.scan +private Scan +AsyncRpcRetryingCallerFactory.ScanSingleRegionCallerBuilder.scan private Scan @@ -339,11 +339,11 @@ service. protected Scan -ClientScanner.getScan() +ScannerCallable.getScan() protected Scan -ScannerCallable.getScan() +ClientScanner.getScan() Scan @@ -638,29 +638,29 @@ service. ResultScanner -AsyncTable.getScanner(Scanscan) -Returns a scanner on the current table as specified by the Scan object. - +RawAsyncTableImpl.getScanner(Scanscan) ResultScanner -Table.getScanner(Scanscan) -Returns a scanner on the current table as specified by the Scan - object. +HTable.getScanner(Scanscan) +The underlying HTable must not be closed. ResultScanner -AsyncTableImpl.getScanner(Scanscan) +Table.getScanner(Scanscan) +Returns a scanner on the current table as specified by the Scan + object. + ResultScanner -RawAsyncTableImpl.getScanner(Scanscan) +AsyncTableImpl.getScanner(Scanscan) ResultScanner -HTable.getScanner(Scanscan) -The underlying HTable must not be closed. +AsyncTable.getScanner(Scanscan) +Returns a scanner on the current table as specified by the Scan object. @@ -703,9 +703,7 @@ service. http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListResult -AsyncTable.scanAll(Scanscan) -Return all the results that match the given scan object. - +RawAsyncTableImpl.scanAll(Scanscan) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListResult @@ -713,7 +711,9 @@ service. http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.html b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.html index ad601c4..53e455f 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.html @@ -1117,1183 +1117,1186 @@ 1109 @Nullable 1110 public static TableState getTableState(Connection conn, TableName tableName) throws IOException { -1112Table metaHTable = getMetaHTable(conn); -1113Get get = new Get(tableName.getName()).addColumn(getTableFamily(), getTableStateColumn()); -1114long time = EnvironmentEdgeManager.currentTime(); -1115get.setTimeRange(0, time); -1116Result result = -1117metaHTable.get(get); -1118return getTableState(result); -1119 } -1120 -1121 /** -1122 * Fetch table states from META table -1123 * @param conn connection to use -1124 * @return map {tableName -gt; state} -1125 * @throws IOException -1126 */ -1127 public static MapTableName, TableState getTableStates(Connection conn) -1128 throws IOException { -1129final MapTableName, TableState states = new LinkedHashMap(); -1130Visitor collector = new Visitor() { -1131 @Override -1132 public boolean visit(Result r) throws IOException { -1133TableState state = getTableState(r); -1134if (state != null) -1135 states.put(state.getTableName(), state); -1136return true; -1137 } -1138}; -1139fullScanTables(conn, collector); -1140return states; -1141 } -1142 -1143 /** -1144 * Updates state in META -1145 * @param conn connection to use -1146 * @param tableName table to look for -1147 * @throws IOException -1148 */ -1149 public static void updateTableState(Connection conn, TableName tableName, -1150 TableState.State actual) throws IOException { -1151updateTableState(conn, new TableState(tableName, actual)); -1152 } -1153 -1154 /** -1155 * Decode table state from META Result. -1156 * Should contain cell from HConstants.TABLE_FAMILY -1157 * @param r result -1158 * @return null if not found -1159 * @throws IOException -1160 */ -1161 @Nullable -1162 public static TableState getTableState(Result r) -1163 throws IOException { -1164Cell cell = r.getColumnLatestCell(getTableFamily(), getTableStateColumn()); -1165if (cell == null) return null; -1166try { -1167 return TableState.parseFrom(TableName.valueOf(r.getRow()), -1168 Arrays.copyOfRange(cell.getValueArray(), -1169 cell.getValueOffset(), cell.getValueOffset() + cell.getValueLength())); -1170} catch (DeserializationException e) { -1171 throw new IOException(e); -1172} -1173 -1174 } -1175 -1176 /** -1177 * Implementations 'visit' a catalog table row. -1178 */ -1179 public interface Visitor { -1180/** -1181 * Visit the catalog table row. -1182 * @param r A row from catalog table -1183 * @return True if we are to proceed scanning the table, else false if -1184 * we are to stop now. -1185 */ -1186boolean visit(final Result r) throws IOException; -1187 } -1188 -1189 /** -1190 * Implementations 'visit' a catalog table row but with close() at the end. -1191 */ -1192 public interface CloseableVisitor extends Visitor, Closeable { -1193 } -1194 -1195 /** -1196 * A {@link Visitor} that collects content out of passed {@link Result}. -1197 */ -1198 static abstract class CollectingVisitorT implements Visitor { -1199final ListT results = new ArrayList(); -1200@Override -1201public boolean visit(Result r) throws IOException { -1202 if (r == null || r.isEmpty()) return true; -1203 add(r); -1204 return true; -1205} -1206 -1207abstract void add(Result r); -1208 -1209/** -1210 * @return Collected results; wait till visits complete to collect all -1211 * possible results -1212 */ -1213ListT getResults() { -1214 return this.results; -1215} -1216 } -1217 -1218 /** -1219 * Collects all returned. -1220 */ -1221 static class CollectAllVisitor extends CollectingVisitorResult { -1222@Override -1223void add(Result r) { -1224 this.results.add(r); -1225} -1226 } -1227 -1228 /** -1229 * A Visitor that skips offline regions and split parents -1230 */ -1231 public static abstract class DefaultVisitorBase implements Visitor { -1232 -1233public DefaultVisitorBase() { -1234 super(); -1235} -1236 -1237public abstract boolean visitInternal(Result rowResult) throws IOException; -1238 -1239@Override -1240public boolean visit(Result rowResult) throws IOException { -1241 RegionInfo info
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1f2eeb22/devapidocs/org/apache/hadoop/hbase/util/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html index ca0e581..7eb86ac 100644 --- a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html @@ -535,14 +535,14 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) +org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.PureJavaComparer (implements org.apache.hadoop.hbase.util.Bytes.ComparerT) org.apache.hadoop.hbase.util.ChecksumType -org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.UnsafeComparer (implements org.apache.hadoop.hbase.util.Bytes.ComparerT) org.apache.hadoop.hbase.util.PrettyPrinter.Unit -org.apache.hadoop.hbase.util.Order -org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.PureJavaComparer (implements org.apache.hadoop.hbase.util.Bytes.ComparerT) org.apache.hadoop.hbase.util.PoolMap.PoolType org.apache.hadoop.hbase.util.IdReadWriteLock.ReferenceType org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE +org.apache.hadoop.hbase.util.Order +org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.UnsafeComparer (implements org.apache.hadoop.hbase.util.Bytes.ComparerT) http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1f2eeb22/devapidocs/org/apache/hadoop/hbase/wal/AbstractFSWALProvider.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/wal/AbstractFSWALProvider.html b/devapidocs/org/apache/hadoop/hbase/wal/AbstractFSWALProvider.html index c08dcaa..e1d8b82 100644 --- a/devapidocs/org/apache/hadoop/hbase/wal/AbstractFSWALProvider.html +++ b/devapidocs/org/apache/hadoop/hbase/wal/AbstractFSWALProvider.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":6,"i2":6,"i3":9,"i4":9,"i5":9,"i6":10,"i7":9,"i8":10,"i9":9,"i10":9,"i11":9,"i12":10,"i13":9,"i14":9,"i15":9,"i16":10,"i17":10,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":10,"i25":9}; +var methods = {"i0":10,"i1":10,"i2":6,"i3":6,"i4":9,"i5":9,"i6":9,"i7":10,"i8":9,"i9":10,"i10":9,"i11":9,"i12":9,"i13":10,"i14":9,"i15":9,"i16":9,"i17":10,"i18":10,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":10,"i26":9}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -283,71 +283,77 @@ implements void +addWALActionsListener(WALActionsListenerlistener) +Add a WALActionsListener. + + + +void close() shutdown utstanding WALs and clean up any persisted state. - + protected abstract T createWAL() - + protected abstract void doInit(org.apache.hadoop.conf.Configurationconf) - + static long extractFileNumFromWAL(WALwal) It returns the file create timestamp from the file name. - + static org.apache.hadoop.fs.Path getArchivedLogPath(org.apache.hadoop.fs.Pathpath, org.apache.hadoop.conf.Configurationconf) Get the archived WAL file path - + static org.apache.hadoop.fs.Path getCurrentFileName(WALwal) return the current filename from the current wal. - + long getLogFileSize() iff the given WALFactory is using the DefaultWALProvider for meta and/or non-meta, count the size of files (only rolled). - + static long getLogFileSize(WALwal) returns the size of rolled WAL files. - + long getNumLogFiles() iff the given WALFactory is using the DefaultWALProvider for meta and/or non-meta, count the number of files (rolled and active). - + static int getNumRolledLogFiles(WALwal) returns the number of rolled WAL files. - + static ServerName getServerNameFromWALDirectoryName(org.apache.hadoop.conf.Configurationconf, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringpath) Pulls a ServerName out of a Path generated according to our layout rules. - + static ServerName getServerNameFromWALDirectoryName(org.apache.hadoop.fs.PathlogFile) This function returns region server name from a log file name which is in one of the following @@ -358,81 +364,79 @@ implements - + T getWAL(RegionInforegion) - + static
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/backup/TestRepairAfterFailedDelete.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/backup/TestRepairAfterFailedDelete.html b/testdevapidocs/org/apache/hadoop/hbase/backup/TestRepairAfterFailedDelete.html index 46c2cc7..416914b 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/backup/TestRepairAfterFailedDelete.html +++ b/testdevapidocs/org/apache/hadoop/hbase/backup/TestRepairAfterFailedDelete.html @@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab"; -public class TestRepairAfterFailedDelete +public class TestRepairAfterFailedDelete extends TestBackupBase @@ -150,6 +150,10 @@ extends Field and Description +static HBaseClassTestRule +CLASS_RULE + + private static org.slf4j.Logger LOG @@ -225,13 +229,22 @@ extends + + + + +CLASS_RULE +public static finalHBaseClassTestRule CLASS_RULE + + LOG -private static finalorg.slf4j.Logger LOG +private static finalorg.slf4j.Logger LOG @@ -248,7 +261,7 @@ extends TestRepairAfterFailedDelete -publicTestRepairAfterFailedDelete() +publicTestRepairAfterFailedDelete() @@ -265,7 +278,7 @@ extends testRepairBackupDelete -publicvoidtestRepairBackupDelete() +publicvoidtestRepairBackupDelete() throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">Exception Throws: http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/backup/TestRestoreBoundaryTests.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/backup/TestRestoreBoundaryTests.html b/testdevapidocs/org/apache/hadoop/hbase/backup/TestRestoreBoundaryTests.html index 0297e64..1267c07 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/backup/TestRestoreBoundaryTests.html +++ b/testdevapidocs/org/apache/hadoop/hbase/backup/TestRestoreBoundaryTests.html @@ -150,6 +150,10 @@ extends Field and Description +static HBaseClassTestRule +CLASS_RULE + + private static org.slf4j.Logger LOG @@ -233,13 +237,22 @@ extends + + + + +CLASS_RULE +public static finalHBaseClassTestRule CLASS_RULE + + LOG -private static finalorg.slf4j.Logger LOG +private static finalorg.slf4j.Logger LOG @@ -273,7 +286,7 @@ extends testFullRestoreSingleEmpty -publicvoidtestFullRestoreSingleEmpty() +publicvoidtestFullRestoreSingleEmpty() throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">Exception Verify that a single empty table is restored to a new table. @@ -288,7 +301,7 @@ extends testFullRestoreMultipleEmpty -publicvoidtestFullRestoreMultipleEmpty() +publicvoidtestFullRestoreMultipleEmpty() throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">Exception Verify that multiple tables are restored to new tables. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/backup/TestSystemTableSnapshot.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/backup/TestSystemTableSnapshot.html b/testdevapidocs/org/apache/hadoop/hbase/backup/TestSystemTableSnapshot.html index 4d45c8f..ef25995 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/backup/TestSystemTableSnapshot.html +++ b/testdevapidocs/org/apache/hadoop/hbase/backup/TestSystemTableSnapshot.html @@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab"; -public class TestSystemTableSnapshot +public class TestSystemTableSnapshot extends TestBackupBase @@ -150,6 +150,10 @@ extends Field and Description +static HBaseClassTestRule +CLASS_RULE + + private static org.slf4j.Logger LOG @@ -227,13 +231,22 @@ extends + + + + +CLASS_RULE +public static finalHBaseClassTestRule CLASS_RULE + + LOG -private static finalorg.slf4j.Logger LOG +private static finalorg.slf4j.Logger LOG @@ -250,7 +263,7 @@ extends TestSystemTableSnapshot -publicTestSystemTableSnapshot() +publicTestSystemTableSnapshot() @@ -267,7 +280,7 @@ extends _testBackupRestoreSystemTable -publicvoid_testBackupRestoreSystemTable() +publicvoid_testBackupRestoreSystemTable() throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">Exception Verify backup system table snapshot. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/backup/example/TestZooKeeperTableArchiveClient.html
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa7ffc92/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.MergeCommand.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.MergeCommand.html b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.MergeCommand.html index eb9e252..667152a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.MergeCommand.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.MergeCommand.html @@ -28,22 +28,22 @@ 020 021import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH; 022import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_BANDWIDTH_DESC; -023import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH; -024import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC; -025import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER; -026import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC; -027import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET; -028import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC; -029import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC; -030import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE; -031import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC; -032import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC; -033import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS; -034import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC; -035import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME; -036import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC; -037import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG; -038import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC; +023import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG; +024import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_DEBUG_DESC; +025import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH; +026import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_PATH_DESC; +027import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER; +028import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_RECORD_NUMBER_DESC; +029import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET; +030import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_BACKUP_DESC; +031import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_SET_DESC; +032import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE; +033import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_DESC; +034import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_TABLE_LIST_DESC; +035import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS; +036import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_WORKERS_DESC; +037import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME; +038import static org.apache.hadoop.hbase.backup.BackupRestoreConstants.OPTION_YARN_QUEUE_NAME_DESC; 039 040import java.io.IOException; 041import java.net.URI; @@ -70,194 +70,194 @@ 062import org.apache.hadoop.hbase.backup.util.BackupUtils; 063import org.apache.hadoop.hbase.client.Connection; 064import org.apache.hadoop.hbase.client.ConnectionFactory; -065import org.apache.hbase.thirdparty.com.google.common.collect.Lists; -066import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -067import org.apache.yetus.audience.InterfaceAudience; -068 -069/** -070 * General backup commands, options and usage messages -071 */ -072 +065import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; +066import org.apache.yetus.audience.InterfaceAudience; +067 +068import org.apache.hbase.thirdparty.com.google.common.collect.Lists; +069 +070/** +071 * General backup commands, options and usage messages +072 */ 073@InterfaceAudience.Private 074public final class BackupCommands { -075 -076 public final static String INCORRECT_USAGE = "Incorrect usage"; -077 -078 public final static String TOP_LEVEL_NOT_ALLOWED = -079 "Top level (root) folder is not allowed to be a backup destination"; -080 -081 public static final String USAGE = "Usage: hbase backup
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/96e5e102/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/ScannerContext.NextState.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/ScannerContext.NextState.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/ScannerContext.NextState.html index 914b1c6..03a0b2a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/ScannerContext.NextState.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/ScannerContext.NextState.html @@ -59,646 +59,711 @@ 051@InterfaceStability.Evolving 052public class ScannerContext { 053 -054 /** -055 * Two sets of the same fields. One for the limits, another for the progress towards those limits -056 */ -057 LimitFields limits; -058 LimitFields progress; -059 -060 /** -061 * The state of the scanner after the invocation of {@link InternalScanner#next(java.util.List)} -062 * or {@link RegionScanner#next(java.util.List)}. -063 */ -064 NextState scannerState; -065 private static final NextState DEFAULT_STATE = NextState.MORE_VALUES; -066 -067 /** -068 * Used as an indication to invocations of {@link InternalScanner#next(java.util.List)} and -069 * {@link RegionScanner#next(java.util.List)} that, if true, the progress tracked within this -070 * {@link ScannerContext} instance should be considered while evaluating the limits. Useful for -071 * enforcing a set of limits across multiple calls (i.e. the limit may not be reached in a single -072 * invocation, but any progress made should be considered in future invocations) -073 * p -074 * Defaulting this value to false means that, by default, any tracked progress will be wiped clean -075 * on invocations to {@link InternalScanner#next(java.util.List)} and -076 * {@link RegionScanner#next(java.util.List)} and the call will be treated as though no progress -077 * has been made towards the limits so far. -078 * p -079 * This is an important mechanism. Users of Internal/Region scanners expect that they can define -080 * some limits and then repeatedly invoke {@link InternalScanner#next(List)} or -081 * {@link RegionScanner#next(List)} where each invocation respects these limits separately. -082 * p -083 * For example: pre {@code -084 * ScannerContext context = new ScannerContext.newBuilder().setBatchLimit(5).build(); -085 * RegionScanner scanner = ... -086 * ListCell results = new ArrayListCell(); -087 * while(scanner.next(results, context)) { -088 * // Do something with a batch of 5 cells -089 * } -090 * }/pre However, in the case of RPCs, the server wants to be able to define a set of -091 * limits for a particular RPC request and have those limits respected across multiple -092 * invocations. This means that the progress made towards the limits in earlier calls will be -093 * saved and considered in future invocations -094 */ -095 boolean keepProgress; -096 private static boolean DEFAULT_KEEP_PROGRESS = false; -097 -098 private Cell lastPeekedCell = null; +054 LimitFields limits; +055 /** +056 * A different set of progress fields. Only include batch, dataSize and heapSize. Compare to +057 * LimitFields, ProgressFields doesn't contain time field. As we save a deadline in LimitFields, +058 * so use {@link System#currentTimeMillis()} directly when check time limit. +059 */ +060 ProgressFields progress; +061 +062 /** +063 * The state of the scanner after the invocation of {@link InternalScanner#next(java.util.List)} +064 * or {@link RegionScanner#next(java.util.List)}. +065 */ +066 NextState scannerState; +067 private static final NextState DEFAULT_STATE = NextState.MORE_VALUES; +068 +069 /** +070 * Used as an indication to invocations of {@link InternalScanner#next(java.util.List)} and +071 * {@link RegionScanner#next(java.util.List)} that, if true, the progress tracked within this +072 * {@link ScannerContext} instance should be considered while evaluating the limits. Useful for +073 * enforcing a set of limits across multiple calls (i.e. the limit may not be reached in a single +074 * invocation, but any progress made should be considered in future invocations) +075 * p +076 * Defaulting this value to false means that, by default, any tracked progress will be wiped clean +077 * on invocations to {@link InternalScanner#next(java.util.List)} and +078 * {@link RegionScanner#next(java.util.List)} and the call will be treated as though no progress +079 * has been made towards the limits so far. +080 * p +081 * This is an important mechanism. Users of Internal/Region scanners expect that they can define +082 * some limits and then repeatedly invoke {@link InternalScanner#next(List)} or +083 * {@link RegionScanner#next(List)} where each invocation respects these limits
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8118541f/devapidocs/org/apache/hadoop/hbase/master/DeadServer.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/DeadServer.html b/devapidocs/org/apache/hadoop/hbase/master/DeadServer.html index d880bf7..3d34dcf 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/DeadServer.html +++ b/devapidocs/org/apache/hadoop/hbase/master/DeadServer.html @@ -332,7 +332,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? ServerNameDeathDateComparator -private statichttp://docs.oracle.com/javase/8/docs/api/java/util/Comparator.html?is-external=true; title="class or interface in java.util">ComparatorPairServerName,http://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">Long ServerNameDeathDateComparator +private statichttp://docs.oracle.com/javase/8/docs/api/java/util/Comparator.html?is-external=true; title="class or interface in java.util">ComparatorPairServerName,http://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">Long ServerNameDeathDateComparator @@ -488,7 +488,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? toString -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringtoString() +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringtoString() Overrides: http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--; title="class or interface in java.lang">toStringin classhttp://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object @@ -501,7 +501,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? copyDeadServersSince -publichttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListPairServerName,http://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">LongcopyDeadServersSince(longts) +publichttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListPairServerName,http://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">LongcopyDeadServersSince(longts) Extract all the servers dead since a given time, and sort them. Parameters: @@ -517,7 +517,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? getTimeOfDeath -publichttp://docs.oracle.com/javase/8/docs/api/java/util/Date.html?is-external=true; title="class or interface in java.util">DategetTimeOfDeath(ServerNamedeadServerName) +publichttp://docs.oracle.com/javase/8/docs/api/java/util/Date.html?is-external=true; title="class or interface in java.util">DategetTimeOfDeath(ServerNamedeadServerName) Get the time when a server died Parameters: @@ -533,7 +533,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? removeDeadServer -publicbooleanremoveDeadServer(ServerNamedeadServerName) +publicbooleanremoveDeadServer(ServerNamedeadServerName) remove the specified dead server Parameters:
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/14db89d7/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/RSGroupListTmplImpl.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/RSGroupListTmplImpl.html b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/RSGroupListTmplImpl.html new file mode 100644 index 000..1f26c87 --- /dev/null +++ b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/RSGroupListTmplImpl.html @@ -0,0 +1,560 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + +Source code + + + + +001// Autogenerated Jamon implementation +002// /home/jenkins/jenkins-slave/workspace/hbase_generate_website/hbase/hbase-server/src/main/jamon/org/apache/hadoop/hbase/tmpl/master/RSGroupListTmpl.jamon +003 +004package org.apache.hadoop.hbase.tmpl.master; +005 +006// 27, 5 +007import java.util.Collections; +008// 28, 5 +009import java.util.List; +010// 29, 5 +011import java.util.Map; +012// 30, 5 +013import java.util.Set; +014// 31, 5 +015import java.util.stream.Collectors; +016// 32, 5 +017import org.apache.hadoop.hbase.master.HMaster; +018// 33, 5 +019import org.apache.hadoop.hbase.ServerLoad; +020// 34, 5 +021import org.apache.hadoop.hbase.RSGroupTableAccessor; +022// 35, 5 +023import org.apache.hadoop.hbase.master.ServerManager; +024// 36, 5 +025import org.apache.hadoop.hbase.net.Address; +026// 37, 5 +027import org.apache.hadoop.hbase.rsgroup.RSGroupInfo; +028// 38, 5 +029import org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix; +030 +031public class RSGroupListTmplImpl +032 extends org.jamon.AbstractTemplateImpl +033 implements org.apache.hadoop.hbase.tmpl.master.RSGroupListTmpl.Intf +034 +035{ +036 private final HMaster master; +037 private final ServerManager serverManager; +038 protected static org.apache.hadoop.hbase.tmpl.master.RSGroupListTmpl.ImplData __jamon_setOptionalArguments(org.apache.hadoop.hbase.tmpl.master.RSGroupListTmpl.ImplData p_implData) +039 { +040return p_implData; +041 } +042 public RSGroupListTmplImpl(org.jamon.TemplateManager p_templateManager, org.apache.hadoop.hbase.tmpl.master.RSGroupListTmpl.ImplData p_implData) +043 { +044super(p_templateManager, __jamon_setOptionalArguments(p_implData)); +045master = p_implData.getMaster(); +046serverManager = p_implData.getServerManager(); +047 } +048 +049 @Override public void renderNoFlush(final java.io.Writer jamonWriter) +050throws java.io.IOException +051 { +052// 40, 1 +053 +054ListRSGroupInfo groups = RSGroupTableAccessor.getAllRSGroupInfo(master.getConnection()); +055 +056// 44, 1 +057if ((groups != null groups.size() 0)) +058{ +059 // 44, 45 +060 jamonWriter.write("\n\n"); +061 // 46, 1 +062 +063RSGroupInfo [] rsGroupInfos = groups.toArray(new RSGroupInfo[groups.size()]); +064MapAddress, ServerLoad collectServers = Collections.emptyMap(); +065if (master.getServerManager() != null) { +066 collectServers = +067 master.getServerManager().getOnlineServers().entrySet().stream() +068 .collect(Collectors.toMap(p - p.getKey().getAddress(), Map.Entry::getValue)); +069} +070 +071 // 56, 1 +072 jamonWriter.write("div class=\"tabbable\"\nul class=\"nav nav-pills\"\nli class=\"active\"a href=\"#tab_rsgroup_baseStats\" data-toggle=\"tab\"Base Stats/a/li\nli class=\"\"a href=\"#tab_rsgroup_memoryStats\" data-toggle=\"tab\"Memory/a/li\nli class=\"\"a href=\"#tab_rsgroup_requestStats\" data-toggle=\"tab\"Requests/a/li\nli class=\"\"a href=\"#tab_rsgroup_storeStats\" data-toggle=\"tab\"Storefiles/a/li\nli class=\"\"a href=\"#tab_rsgroup_compactStats\" data-toggle=\"tab\"Compactions/a/li\n/ul\n div class=\"tab-content\" style=\"padding-bottom: 9px; border-bottom: 1px solid #ddd;\"\ndiv class=\"tab-pane active\" id=\"tab_rsgroup_baseStats\"\n"); +073 // 66, 13 +074 { +075// 66, 13 +076 __jamon_innerUnit__rsgroup_baseStats(jamonWriter, rsGroupInfos, collectServers ); +077 } +078 // 66, 97 +079 jamonWriter.write("\n /div\ndiv class=\"tab-pane\" id=\"tab_rsgroup_memoryStats\"\n"); +080 // 69, 13 +081 { +082// 69, 13 +083 __jamon_innerUnit__rsgroup_memoryStats(jamonWriter, rsGroupInfos, collectServers ); +084 } +085 // 69, 99 +086 jamonWriter.write("\n /div\ndiv class=\"tab-pane\" id=\"tab_rsgroup_requestStats\"\n"); +087 // 72, 13 +088 { +089// 72, 13 +090 __jamon_innerUnit__rsgroup_requestStats(jamonWriter, rsGroupInfos, collectServers ); +091 } +092 // 72, 100 +093 jamonWriter.write("\n /div\ndiv class=\"tab-pane\" id=\"tab_rsgroup_storeStats\"\n"); +094 // 75, 13 +095 { +096// 75, 13
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/testdevapidocs/org/apache/hadoop/hbase/TestMovedRegionsCleaner.TestMockRegionServer.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/TestMovedRegionsCleaner.TestMockRegionServer.html b/testdevapidocs/org/apache/hadoop/hbase/TestMovedRegionsCleaner.TestMockRegionServer.html index 0048578..d5b9f50 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/TestMovedRegionsCleaner.TestMockRegionServer.html +++ b/testdevapidocs/org/apache/hadoop/hbase/TestMovedRegionsCleaner.TestMockRegionServer.html @@ -180,7 +180,7 @@ extends Methods inherited from classorg.apache.hadoop.hbase.regionserver.HRegionServer -abort, addRegion, addToMovedRegions, canCreateBaseZNode, canUpdateTableDescriptor, checkFileSystem, clearRegionBlockCache, closeAllRegions, closeAndOfflineRegionForSplitOrMerge, closeRegion, configureInfoServer, constructRegionServer, convertThrowableToIOE, createClusterConnection, createConnection, createRegionLoad, createRegionServerStatusStub, createRegionServerStatusStub, createRpcServices, execRegionServerService, executeProcedure, getCacheConfig, getChoreService, getClusterConnection, getClusterId, getCompactionPressure, getCompactionRequestor, getCompactSplitThread, getConfiguration, getConfigurationManager, getConnection, getCoordinatedStateManager, getDumpServlet, getEventLoopGroupConfig, getExecutorService, getFavoredNodesForRegion, getFileSystem, getFlushPressure, getFlushRequester, getFlushThroughputController, getFsTableDescriptors, getHeapMemoryManager, getInfoServer, getLastSequenceId, getLeases, getMasterAddressTracker, getMetaTableLocator, getMetaTableObserver , getMetrics, getMostLoadedRegions, getNonceManager, getNumberOfOnlineRegions, getOnlineRegion, getOnlineRegionsLocalContext, getOnlineTables, getProcessName, getRegion, getRegion, getRegionBlockLocations, getRegionByEncodedName, getRegionByEncodedName, getRegions, getRegions, getRegionServerAccounting, getRegionServerCoprocessorHost, getRegionServerCoprocessors, getRegionServerMetrics, getRegionServerRpcQuotaManager, getRegionServerSpaceQuotaManager, getRegionsInTransitionInRS, getReplicationSourceService, getRootDir, getRpcServer, getRSRpcServices, getSecureBulkLoadManager, getServerName, getStartcode, getTableDescriptors, getThreadWakeFrequency, getWAL, getWALFileSystem, getWalRoller, getWALRootDir, getWALs, getZooKeeper, initializeMemStoreChunkCreator, isAborted, isOnline, isStopped, isStopping, login, main, onConfigurationChange, postOpenDeployTasks, regionLock, registerService, remoteProcedureComplete, removeRegion, reportRegionSizesForQuotas, reportRegionStateTransition, send ShutdownInterrupt, setInitLatch, setupClusterConnection, shouldUseThisHostnameInstead, stop, stop, stopServiceThreads, toString, tryRegionServerReport, unassign, updateConfiguration, updateRegionFavoredNodesMapping, waitForMasterActive, waitForServerOnline, walRollRequestFinished +abort, addRegion, addToMovedRegions, canCreateBaseZNode, canUpdateTableDescriptor, checkFileSystem, clearRegionBlockCache, closeAllRegions, closeAndOfflineRegionForSplitOrMerge, closeRegion, configureInfoServer, constructRegionServer, convertThrowableToIOE, createClusterConnection, createConnection, createRegionLoad, createRegionServerStatusStub, createRegionServerStatusStub, createRpcServices, execRegionServerService, executeProcedure, getCacheConfig, getChoreService, getClusterConnection, getClusterId, getCompactionPressure, getCompactionRequestor, getCompactSplitThread, getConfiguration, getConfigurationManager, getConnection, getCoordinatedStateManager, getDumpServlet, getEventLoopGroupConfig, getExecutorService, getFavoredNodesForRegion, getFileSystem, getFlushPressure, getFlushRequester, getFlushThroughputController, getFsTableDescriptors, getHeapMemoryManager, getInfoServer, getLastSequenceId, getLeases, getMasterAddressTracker, getMetaTableLocator, getMetaTableObserver , getMetrics, getMostLoadedRegions, getNonceManager, getNumberOfOnlineRegions, getOnlineRegion, getOnlineRegionsLocalContext, getOnlineTables, getProcessName, getRegion, getRegion, getRegionBlockLocations, getRegionByEncodedName, getRegionByEncodedName, getRegions, getRegions, getRegionServerAccounting, getRegionServerCoprocessorHost, getRegionServerCoprocessors, getRegionServerMetrics, getRegionServerRpcQuotaManager, getRegionServerSpaceQuotaManager, getRegionsInTransitionInRS, getReplicationSourceService, getRootDir, getRpcServer, getRSRpcServices, getSecureBulkLoadManager, getServerName, getStartcode, getTableDescriptors, getThreadWakeFrequency, getWAL, getWALFileSystem, getWalRoller, getWALRootDir, getWALs, getZooKeeper, initializeMemStoreChunkCreator, isAborted, isOnline, isStopped, isStopping, login, main, onConfigurationChange, postOpenDeployTasks, regionLock, registerService,
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f183e80f/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html b/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html index fadf667..14b2b69 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html @@ -7,269 +7,269 @@ 001/** -002 * -003 * Licensed to the Apache Software Foundation (ASF) under one -004 * or more contributor license agreements. See the NOTICE file -005 * distributed with this work for additional information -006 * regarding copyright ownership. The ASF licenses this file -007 * to you under the Apache License, Version 2.0 (the -008 * "License"); you may not use this file except in compliance -009 * with the License. You may obtain a copy of the License at -010 * -011 * http://www.apache.org/licenses/LICENSE-2.0 -012 * -013 * Unless required by applicable law or agreed to in writing, software -014 * distributed under the License is distributed on an "AS IS" BASIS, -015 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -016 * See the License for the specific language governing permissions and -017 * limitations under the License. -018 */ +002 * Licensed to the Apache Software Foundation (ASF) under one +003 * or more contributor license agreements. See the NOTICE file +004 * distributed with this work for additional information +005 * regarding copyright ownership. The ASF licenses this file +006 * to you under the Apache License, Version 2.0 (the +007 * "License"); you may not use this file except in compliance +008 * with the License. You may obtain a copy of the License at +009 * +010 * http://www.apache.org/licenses/LICENSE-2.0 +011 * +012 * Unless required by applicable law or agreed to in writing, software +013 * distributed under the License is distributed on an "AS IS" BASIS, +014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +015 * See the License for the specific language governing permissions and +016 * limitations under the License. +017 */ +018package org.apache.hadoop.hbase.wal; 019 -020 -021package org.apache.hadoop.hbase.wal; -022 -023import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; -024 -025import java.io.IOException; -026import java.io.InterruptedIOException; -027import java.util.Collections; -028import java.util.List; -029import java.util.OptionalLong; -030import java.util.concurrent.atomic.AtomicReference; -031 -032import org.apache.hadoop.conf.Configuration; -033import org.apache.hadoop.fs.FileSystem; -034import org.apache.hadoop.fs.Path; -035import org.apache.yetus.audience.InterfaceAudience; -036import org.slf4j.Logger; -037import org.slf4j.LoggerFactory; -038// imports for things that haven't moved from regionserver.wal yet. -039import org.apache.hadoop.hbase.regionserver.wal.MetricsWAL; -040import org.apache.hadoop.hbase.regionserver.wal.ProtobufLogReader; -041import org.apache.hadoop.hbase.regionserver.wal.WALActionsListener; -042import org.apache.hadoop.hbase.replication.regionserver.WALFileLengthProvider; -043import org.apache.hadoop.hbase.util.CancelableProgressable; -044import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -045import org.apache.hadoop.hbase.util.LeaseNotRecoveredException; -046import org.apache.hadoop.hbase.wal.WAL.Reader; -047import org.apache.hadoop.hbase.wal.WALProvider.Writer; -048 -049/** -050 * Entry point for users of the Write Ahead Log. -051 * Acts as the shim between internal use and the particular WALProvider we use to handle wal -052 * requests. -053 * -054 * Configure which provider gets used with the configuration setting "hbase.wal.provider". Available -055 * implementations: -056 * ul -057 * liemdefaultProvider/em : whatever provider is standard for the hbase version. Currently -058 * "filesystem"/li -059 * liemfilesystem/em : a provider that will run on top of an implementation of the Hadoop -060 * FileSystem interface, normally HDFS./li -061 * liemmultiwal/em : a provider that will use multiple "filesystem" wal instances per region -062 * server./li -063 * /ul -064 * -065 * Alternatively, you may provide a custom implementation of {@link WALProvider} by class name. -066 */ -067@InterfaceAudience.Private -068public class WALFactory implements WALFileLengthProvider { -069 -070 private static final Logger LOG = LoggerFactory.getLogger(WALFactory.class); -071 -072 /** -073 * Maps between configuration names for providers and implementation classes. -074 */ -075 static enum Providers { -076
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer.html b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer.html index 685363b..4a7ef76 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer.html +++ b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer.html @@ -49,7 +49,7 @@ var activeTableTab = "activeTableTab"; -PrevClass +PrevClass NextClass @@ -127,7 +127,7 @@ var activeTableTab = "activeTableTab"; -private class RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer +private class RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer extends RawAsyncHBaseAdmin.TableProcedureBiConsumer @@ -232,7 +232,7 @@ extends SplitTableRegionProcedureBiConsumer -SplitTableRegionProcedureBiConsumer(TableNametableName) +SplitTableRegionProcedureBiConsumer(TableNametableName) @@ -249,7 +249,7 @@ extends getOperationType -http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetOperationType() +http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetOperationType() Specified by: getOperationTypein classRawAsyncHBaseAdmin.TableProcedureBiConsumer @@ -284,7 +284,7 @@ extends -PrevClass +PrevClass NextClass http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html index bd1c256..98a34ec 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html +++ b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html @@ -110,7 +110,7 @@ var activeTableTab = "activeTableTab"; http://docs.oracle.com/javase/8/docs/api/java/lang/FunctionalInterface.html?is-external=true; title="class or interface in java.lang">@FunctionalInterface -private static interface RawAsyncHBaseAdmin.TableOperator +private static interface RawAsyncHBaseAdmin.TableOperator @@ -154,7 +154,7 @@ private static interface operate -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Voidoperate(TableNametable) +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Voidoperate(TableNametable) http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableProcedureBiConsumer.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableProcedureBiConsumer.html b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableProcedureBiConsumer.html index af56e24..d31719c 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableProcedureBiConsumer.html +++ b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableProcedureBiConsumer.html @@ -126,7 +126,7 @@ var activeTableTab = "activeTableTab"; -private abstract class RawAsyncHBaseAdmin.TableProcedureBiConsumer +private abstract class RawAsyncHBaseAdmin.TableProcedureBiConsumer extends RawAsyncHBaseAdmin.ProcedureBiConsumer @@ -240,7 +240,7 @@ extends tableName -protected finalTableName tableName +protected finalTableName tableName @@ -257,7 +257,7 @@ extends TableProcedureBiConsumer -TableProcedureBiConsumer(TableNametableName) +TableProcedureBiConsumer(TableNametableName) @@ -274,7 +274,7 @@ extends getOperationType -abstracthttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetOperationType() +abstracthttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetOperationType() @@ -283,7 +283,7 @@ extends getDescription
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c7c40c62/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html b/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html index 429ebaa..b4404ce 100644 --- a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html +++ b/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html @@ -612,8 +612,8 @@ default void -MasterObserver.postGetClusterStatus(ObserverContextMasterCoprocessorEnvironmentctx, -ClusterStatusstatus) +MasterObserver.postGetClusterMetrics(ObserverContextMasterCoprocessorEnvironmentctx, + ClusterMetricsstatus) Called after get cluster status. @@ -1452,7 +1452,7 @@ default void -MasterObserver.preGetClusterStatus(ObserverContextMasterCoprocessorEnvironmentctx) +MasterObserver.preGetClusterMetrics(ObserverContextMasterCoprocessorEnvironmentctx) Called before get cluster status. @@ -2678,166 +2678,162 @@ void -AccessController.preGetClusterStatus(ObserverContextMasterCoprocessorEnvironmentctx) - - -void AccessController.preGetLocks(ObserverContextMasterCoprocessorEnvironmentctx) - + void AccessController.preGetNamespaceDescriptor(ObserverContextMasterCoprocessorEnvironmentctx, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringnamespace) - + void AccessController.preGetOp(ObserverContextRegionCoprocessorEnvironmentc, Getget, http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListCellresult) - + void AccessController.preGetProcedures(ObserverContextMasterCoprocessorEnvironmentctx) - + void AccessController.preGetReplicationPeerConfig(ObserverContextMasterCoprocessorEnvironmentctx, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringpeerId) - + void AccessController.preGetTableDescriptors(ObserverContextMasterCoprocessorEnvironmentctx, http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableNametableNamesList, http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableDescriptordescriptors, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringregex) - + Result AccessController.preIncrement(ObserverContextRegionCoprocessorEnvironmentc, Incrementincrement) - + Result AccessController.preIncrementAfterRowLock(ObserverContextRegionCoprocessorEnvironmentc, Incrementincrement) - + void AccessController.preListDecommissionedRegionServers(ObserverContextMasterCoprocessorEnvironmentctx) - + void AccessController.preListReplicationPeers(ObserverContextMasterCoprocessorEnvironmentctx, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringregex) - + void AccessController.preListSnapshot(ObserverContextMasterCoprocessorEnvironmentctx, SnapshotDescriptionsnapshot) - + void AccessController.preLockHeartbeat(ObserverContextMasterCoprocessorEnvironmentctx, TableNametableName, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringdescription) - + void AccessController.preMergeRegions(ObserverContextMasterCoprocessorEnvironmentctx, RegionInfo[]regionsToMerge) - + void AccessController.preModifyNamespace(ObserverContextMasterCoprocessorEnvironmentctx, NamespaceDescriptorns) - + void AccessController.preModifyTable(ObserverContextMasterCoprocessorEnvironmentc, TableNametableName, TableDescriptorhtd) - + void CoprocessorWhitelistMasterObserver.preModifyTable(ObserverContextMasterCoprocessorEnvironmentctx, TableNametableName, TableDescriptorhtd) - + void AccessController.preMove(ObserverContextMasterCoprocessorEnvironmentc, RegionInforegion, ServerNamesrcServer, ServerNamedestServer) - + void AccessController.preMoveServers(ObserverContextMasterCoprocessorEnvironmentctx, http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">SetAddressservers,
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bb398572/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html index 5b3b750..a1f3f7e 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html @@ -97,3307 +97,3304 @@ 089import org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer; 090import org.apache.hbase.thirdparty.io.netty.util.Timeout; 091import org.apache.hbase.thirdparty.io.netty.util.TimerTask; -092import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -093import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter; -094import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService; -095import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest; -096import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesResponse; -097import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest; -098import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse; -099import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest; -100import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionResponse; -101import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionRequest; -102import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionResponse; -103import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest; -104import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse; -105import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequest; -106import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse; -107import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest; -108import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse; -109import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest; -110import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerResponse; -111import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest; -112import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationResponse; -113import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription; -114import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType; -115import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema; -116import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest; -117import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse; -118import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest; -119import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse; -120import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest; -121import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse; -122import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest; -123import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse; -124import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest; -125import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse; -126import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest; -127import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse; -128import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest; -129import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse; -130import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest; -131import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse; -132import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest; -133import
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.Callback.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.Callback.html b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.Callback.html index c8dbf3a..1116704 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.Callback.html +++ b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.Callback.html @@ -248,6 +248,6 @@ public static interface Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.html b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.html index c89b6ca..721a4d2 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.html +++ b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/Batch.html @@ -264,6 +264,6 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? -Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/coprocessor/BigDecimalColumnInterpreter.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/BigDecimalColumnInterpreter.html b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/BigDecimalColumnInterpreter.html index 280f7cf..5a38ff2 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/BigDecimalColumnInterpreter.html +++ b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/BigDecimalColumnInterpreter.html @@ -649,6 +649,6 @@ extends Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/coprocessor/DoubleColumnInterpreter.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/DoubleColumnInterpreter.html b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/DoubleColumnInterpreter.html index 65aa67d..c0888a4 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/DoubleColumnInterpreter.html +++ b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/DoubleColumnInterpreter.html @@ -640,6 +640,6 @@ extends Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.html b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.html index 2a6aa03..9bc8eff 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.html +++ b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.html @@ -640,6 +640,6 @@ extends Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/client/coprocessor/RowProcessorClient.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/RowProcessorClient.html b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/RowProcessorClient.html index 07901b7..ff53947 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/coprocessor/RowProcessorClient.html +++ b/devapidocs/org/apache/hadoop/hbase/client/coprocessor/RowProcessorClient.html @@ -276,6 +276,6 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? -Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072018
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/83bf6175/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html index 985778f..854ba52 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html @@ -662,1932 +662,1924 @@ 654 completeCompaction(toBeRemovedStoreFiles); 655 } 656 -657 private HStoreFile createStoreFileAndReader(final Path p) throws IOException { -658StoreFileInfo info = new StoreFileInfo(conf, this.getFileSystem(), p); -659return createStoreFileAndReader(info); -660 } -661 -662 private HStoreFile createStoreFileAndReader(StoreFileInfo info) throws IOException { -663 info.setRegionCoprocessorHost(this.region.getCoprocessorHost()); -664HStoreFile storeFile = new HStoreFile(this.getFileSystem(), info, this.conf, this.cacheConf, -665this.family.getBloomFilterType(), isPrimaryReplicaStore()); -666storeFile.initReader(); -667return storeFile; -668 } -669 -670 /** -671 * This message intends to inform the MemStore that next coming updates -672 * are going to be part of the replaying edits from WAL -673 */ -674 public void startReplayingFromWAL(){ -675 this.memstore.startReplayingFromWAL(); -676 } -677 -678 /** -679 * This message intends to inform the MemStore that the replaying edits from WAL -680 * are done -681 */ -682 public void stopReplayingFromWAL(){ -683 this.memstore.stopReplayingFromWAL(); -684 } -685 -686 /** -687 * Adds a value to the memstore -688 */ -689 public void add(final Cell cell, MemStoreSizing memstoreSizing) { -690lock.readLock().lock(); -691try { -692 this.memstore.add(cell, memstoreSizing); -693} finally { -694 lock.readLock().unlock(); -695} -696 } -697 -698 /** -699 * Adds the specified value to the memstore -700 */ -701 public void add(final IterableCell cells, MemStoreSizing memstoreSizing) { -702lock.readLock().lock(); -703try { -704 memstore.add(cells, memstoreSizing); -705} finally { -706 lock.readLock().unlock(); -707} -708 } -709 -710 @Override -711 public long timeOfOldestEdit() { -712return memstore.timeOfOldestEdit(); -713 } -714 -715 /** -716 * @return All store files. -717 */ -718 @Override -719 public CollectionHStoreFile getStorefiles() { -720return this.storeEngine.getStoreFileManager().getStorefiles(); -721 } -722 -723 @Override -724 public CollectionHStoreFile getCompactedFiles() { -725return this.storeEngine.getStoreFileManager().getCompactedfiles(); -726 } -727 -728 /** -729 * This throws a WrongRegionException if the HFile does not fit in this region, or an -730 * InvalidHFileException if the HFile is not valid. -731 */ -732 public void assertBulkLoadHFileOk(Path srcPath) throws IOException { -733HFile.Reader reader = null; -734try { -735 LOG.info("Validating hfile at " + srcPath + " for inclusion in " -736 + "store " + this + " region " + this.getRegionInfo().getRegionNameAsString()); -737 reader = HFile.createReader(srcPath.getFileSystem(conf), srcPath, cacheConf, -738isPrimaryReplicaStore(), conf); -739 reader.loadFileInfo(); -740 -741 Optionalbyte[] firstKey = reader.getFirstRowKey(); -742 Preconditions.checkState(firstKey.isPresent(), "First key can not be null"); -743 OptionalCell lk = reader.getLastKey(); -744 Preconditions.checkState(lk.isPresent(), "Last key can not be null"); -745 byte[] lastKey = CellUtil.cloneRow(lk.get()); -746 -747 LOG.debug("HFile bounds: first=" + Bytes.toStringBinary(firstKey.get()) + -748 " last=" + Bytes.toStringBinary(lastKey)); -749 LOG.debug("Region bounds: first=" + -750 Bytes.toStringBinary(getRegionInfo().getStartKey()) + -751 " last=" + Bytes.toStringBinary(getRegionInfo().getEndKey())); -752 -753 if (!this.getRegionInfo().containsRange(firstKey.get(), lastKey)) { -754throw new WrongRegionException( -755"Bulk load file " + srcPath.toString() + " does not fit inside region " -756+ this.getRegionInfo().getRegionNameAsString()); -757 } -758 -759 if(reader.length() conf.getLong(HConstants.HREGION_MAX_FILESIZE, -760 HConstants.DEFAULT_MAX_FILE_SIZE)) { -761LOG.warn("Trying to bulk load hfile " + srcPath.toString() + " with size: " + -762reader.length() + " bytes can be problematic as it may lead to oversplitting."); -763 } -764 -765 if (verifyBulkLoads) { -766long verificationStartTime = EnvironmentEdgeManager.currentTime(); -767LOG.info("Full verification started
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/devapidocs/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.html b/devapidocs/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.html index 7793de4..b03b542 100644 --- a/devapidocs/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.html +++ b/devapidocs/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.html @@ -189,7 +189,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? Field and Description -private static org.apache.hadoop.hbase.shaded.io.netty.buffer.ByteBufAllocator +private static org.apache.hbase.thirdparty.io.netty.buffer.ByteBufAllocator ALLOC @@ -293,8 +293,8 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? longfileId) -private static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.hbase.shaded.io.netty.util.concurrent.Futureorg.apache.hadoop.hbase.shaded.io.netty.channel.Channel -connectToDataNodes(org.apache.hadoop.conf.Configurationconf, +private static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hbase.thirdparty.io.netty.util.concurrent.Futureorg.apache.hbase.thirdparty.io.netty.channel.Channel +connectToDataNodes(org.apache.hadoop.conf.Configurationconf, org.apache.hadoop.hdfs.DFSClientclient, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringclientName, org.apache.hadoop.hdfs.protocol.LocatedBlocklocatedBlock, @@ -302,8 +302,8 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? longlatestGS, org.apache.hadoop.hdfs.protocol.datatransfer.BlockConstructionStagestage, org.apache.hadoop.util.DataChecksumsummer, - org.apache.hadoop.hbase.shaded.io.netty.channel.EventLoopGroupeventLoopGroup, - http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">Class? extends org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelchannelClass) + org.apache.hbase.thirdparty.io.netty.channel.EventLoopGroupeventLoopGroup, + http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">Class? extends org.apache.hbase.thirdparty.io.netty.channel.ChannelchannelClass) private static FanOutOneBlockAsyncDFSOutputHelper.BlockAdder @@ -349,27 +349,27 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? static FanOutOneBlockAsyncDFSOutput -createOutput(org.apache.hadoop.hdfs.DistributedFileSystemdfs, +createOutput(org.apache.hadoop.hdfs.DistributedFileSystemdfs, org.apache.hadoop.fs.Pathf, booleanoverwrite, booleancreateParent, shortreplication, longblockSize, - org.apache.hadoop.hbase.shaded.io.netty.channel.EventLoopGroupeventLoopGroup, -http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">Class? extends org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelchannelClass) + org.apache.hbase.thirdparty.io.netty.channel.EventLoopGroupeventLoopGroup, +http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">Class? extends org.apache.hbase.thirdparty.io.netty.channel.ChannelchannelClass) Create a FanOutOneBlockAsyncDFSOutput. private static FanOutOneBlockAsyncDFSOutput -createOutput(org.apache.hadoop.hdfs.DistributedFileSystemdfs, +createOutput(org.apache.hadoop.hdfs.DistributedFileSystemdfs, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringsrc, booleanoverwrite, booleancreateParent, shortreplication, longblockSize, - org.apache.hadoop.hbase.shaded.io.netty.channel.EventLoopGroupeventLoopGroup, -http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or interface in java.lang">Class? extends org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelchannelClass) + org.apache.hbase.thirdparty.io.netty.channel.EventLoopGroupeventLoopGroup, +http://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true; title="class or
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d449e87f/devapidocs/org/apache/hadoop/hbase/RawCell.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/RawCell.html b/devapidocs/org/apache/hadoop/hbase/RawCell.html index a9b9734..0c030f3 100644 --- a/devapidocs/org/apache/hadoop/hbase/RawCell.html +++ b/devapidocs/org/apache/hadoop/hbase/RawCell.html @@ -109,7 +109,7 @@ var activeTableTab = "activeTableTab"; All Known Implementing Classes: -BufferedDataBlockEncoder.OffheapDecodedCell, BufferedDataBlockEncoder.OnheapDecodedCell, ByteBufferChunkCell, ByteBufferKeyValue, IndividualBytesFieldCell, KeyValue, KeyValue.KeyOnlyKeyValue, MapReduceCell, Mutation.CellWrapper, NoTagByteBufferChunkCell, NoTagsByteBufferKeyValue, NoTagsKeyValue, Priv ateCellUtil.EmptyByteBufferCell, PrivateCellUtil.EmptyCell, PrivateCellUtil.FirstOnRowByteBufferCell, PrivateCellUtil.FirstOnRowCell, PrivateCellUtil.FirstOnRowColByteBufferCell, PrivateCellUtil.FirstOnRowColCell, PrivateCellUtil.FirstOnRowColTSByteBufferCell, PrivateCellUtil.FirstOnRowColTSCell, PrivateCellUtil.FirstOnRowDeleteFamilyCell, PrivateCellUtil.LastOnRowByteBufferCell, PrivateCellUtil.LastOnRowCell, PrivateCellUtil.LastOnRowColByteBufferCell, PrivateCellUtil.LastOnRowColCell, PrivateCellUtil.TagRewriteByteBufferCell, PrivateCellUtil.TagRewriteCell, PrivateCellUtil.ValueAndTagRewriteByteBufferCell, PrivateCellUtil.ValueAndTagRewriteCell, SizeCachedKeyValue, SizeCachedNoTagsKeyValue +BufferedDataBlockEncoder.OffheapDecodedExtendedCell, BufferedDataBlockEncoder.OnheapDecodedCell, ByteBufferChunkKeyValue, ByteBufferExtendedCell, ByteBufferKeyOnlyKeyValue, ByteBufferKeyValue, IndividualBytesFieldCell, KeyOnlyFilter.KeyOnlyByteBufferExtendedCell, KeyValue, KeyValue.KeyOnlyKeyValue, MapReduceExtendedCell, Mutation.CellWrapper, NoTagByteBufferChunkKeyValue, NoTagsByteBufferKeyValue, NoTagsKeyValue, PrivateCellUtil.EmptyByteBufferExtendedCell, PrivateCellUtil.EmptyCell, PrivateCellUtil.FirstOnRowByteBufferExtendedCell, PrivateCellUtil.FirstOnRowCell, PrivateCellUtil.FirstOnRowColByteBufferExtendedCell, PrivateCellUtil.FirstOnRowColCell, PrivateCellUtil.FirstOnRowColTSByteBufferExtendedCell, PrivateCellUtil.FirstOnRowColTSCell, PrivateCellUtil.FirstOnR owDeleteFamilyCell, PrivateCellUtil.LastOnRowByteBufferExtendedCell, PrivateCellUtil.LastOnRowCell, PrivateCellUtil.LastOnRowColByteBufferExtendedCell, PrivateCellUtil.LastOnRowColCell, PrivateCellUtil.TagRewriteByteBufferExtendedCell, PrivateCellUtil.TagRewriteCell, PrivateCellUtil.ValueAndTagRewriteByteBufferExtendedCell, PrivateCellUtil.ValueAndTagRewriteCell, SizeCachedKeyValue, SizeCachedNoTagsKeyValue @@ -134,7 +134,7 @@ extends Cell -Cell.DataType +Cell.Type http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d449e87f/devapidocs/org/apache/hadoop/hbase/RawCellBuilder.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/RawCellBuilder.html b/devapidocs/org/apache/hadoop/hbase/RawCellBuilder.html index cc4b37f..d133996 100644 --- a/devapidocs/org/apache/hadoop/hbase/RawCellBuilder.html +++ b/devapidocs/org/apache/hadoop/hbase/RawCellBuilder.html @@ -185,7 +185,7 @@ extends RawCellBuilder -setType(Cell.DataTypetype) +setType(Cell.Typetype) RawCellBuilder @@ -309,16 +309,16 @@ extends + setType -RawCellBuildersetType(Cell.DataTypetype) +RawCellBuildersetType(Cell.Typetype) Specified by: -setTypein interfaceCellBuilder +setTypein interfaceCellBuilder http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d449e87f/devapidocs/org/apache/hadoop/hbase/RegionTooBusyException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/RegionTooBusyException.html b/devapidocs/org/apache/hadoop/hbase/RegionTooBusyException.html index 3c48a21..3f4a0f7 100644 --- a/devapidocs/org/apache/hadoop/hbase/RegionTooBusyException.html +++ b/devapidocs/org/apache/hadoop/hbase/RegionTooBusyException.html @@ -123,11 +123,13 @@ @InterfaceAudience.Public -public class RegionTooBusyException +public class RegionTooBusyException extends http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Thrown by a region server if it will block and wait to serve a request. For example, the client wants to insert something to a region while the - region is compacting. + region is compacting. Keep variance in the passed 'msg' low because its msg is used as a key + over in RetriesExhaustedWithDetailsException + grouping failure types. See Also: Serialized Form @@ -169,14 +171,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/io/IOException.ht Constructor and Description -RegionTooBusyException() -default constructor - - -
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d2b28a1a/devapidocs/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html b/devapidocs/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html index 9d0bd6f..afc9a5b 100644 --- a/devapidocs/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html +++ b/devapidocs/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html @@ -475,7 +475,7 @@ extends org.apache.hadoop.mapreduce.lib.output.FileOutputFormat COMPRESSION_FAMILIES_CONF_KEY -static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String COMPRESSION_FAMILIES_CONF_KEY +static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String COMPRESSION_FAMILIES_CONF_KEY See Also: Constant Field Values @@ -488,7 +488,7 @@ extends org.apache.hadoop.mapreduce.lib.output.FileOutputFormat BLOOM_TYPE_FAMILIES_CONF_KEY -static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String BLOOM_TYPE_FAMILIES_CONF_KEY +static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String BLOOM_TYPE_FAMILIES_CONF_KEY See Also: Constant Field Values @@ -501,7 +501,7 @@ extends org.apache.hadoop.mapreduce.lib.output.FileOutputFormat BLOCK_SIZE_FAMILIES_CONF_KEY -static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String BLOCK_SIZE_FAMILIES_CONF_KEY +static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String BLOCK_SIZE_FAMILIES_CONF_KEY See Also: Constant Field Values @@ -514,7 +514,7 @@ extends org.apache.hadoop.mapreduce.lib.output.FileOutputFormat DATABLOCK_ENCODING_FAMILIES_CONF_KEY -static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String DATABLOCK_ENCODING_FAMILIES_CONF_KEY +static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String DATABLOCK_ENCODING_FAMILIES_CONF_KEY See Also: Constant Field Values @@ -527,7 +527,7 @@ extends org.apache.hadoop.mapreduce.lib.output.FileOutputFormat DATABLOCK_ENCODING_OVERRIDE_CONF_KEY -public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String DATABLOCK_ENCODING_OVERRIDE_CONF_KEY +public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String DATABLOCK_ENCODING_OVERRIDE_CONF_KEY See Also: Constant Field Values @@ -540,7 +540,7 @@ extends org.apache.hadoop.mapreduce.lib.output.FileOutputFormat LOCALITY_SENSITIVE_CONF_KEY -public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String LOCALITY_SENSITIVE_CONF_KEY +public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String LOCALITY_SENSITIVE_CONF_KEY Keep locality while generating HFiles for bulkload. See HBASE-12596 See Also: @@ -554,7 +554,7 @@ extends org.apache.hadoop.mapreduce.lib.output.FileOutputFormat DEFAULT_LOCALITY_SENSITIVE -private static finalboolean DEFAULT_LOCALITY_SENSITIVE +private static finalboolean DEFAULT_LOCALITY_SENSITIVE See Also: Constant Field Values @@ -567,7 +567,7 @@ extends org.apache.hadoop.mapreduce.lib.output.FileOutputFormat OUTPUT_TABLE_NAME_CONF_KEY -static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String OUTPUT_TABLE_NAME_CONF_KEY +static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String OUTPUT_TABLE_NAME_CONF_KEY See Also: Constant Field Values @@ -580,7 +580,7 @@ extends org.apache.hadoop.mapreduce.lib.output.FileOutputFormat MULTI_TABLE_HFILEOUTPUTFORMAT_CONF_KEY -static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String MULTI_TABLE_HFILEOUTPUTFORMAT_CONF_KEY +static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String MULTI_TABLE_HFILEOUTPUTFORMAT_CONF_KEY See Also: Constant Field Values @@ -593,7 +593,7 @@ extends
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b618ac40/devapidocs/org/apache/hadoop/hbase/replication/ReplicationPeerConfigBuilder.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/replication/ReplicationPeerConfigBuilder.html b/devapidocs/org/apache/hadoop/hbase/replication/ReplicationPeerConfigBuilder.html index 71630ed..7a5b5b6 100644 --- a/devapidocs/org/apache/hadoop/hbase/replication/ReplicationPeerConfigBuilder.html +++ b/devapidocs/org/apache/hadoop/hbase/replication/ReplicationPeerConfigBuilder.html @@ -18,8 +18,8 @@ catch(err) { } //--> -var methods = {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6}; -var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"]}; +var methods = {"i0":6,"i1":18,"i2":18,"i3":6,"i4":6,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],16:["t5","Default Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; var tableTab = "tableTab"; @@ -121,7 +121,7 @@ public interface -All MethodsInstance MethodsAbstract Methods +All MethodsInstance MethodsAbstract MethodsDefault Methods Modifier and Type Method and Description @@ -131,47 +131,57 @@ public interface build() -ReplicationPeerConfigBuilder -setBandwidth(longbandwidth) +default ReplicationPeerConfigBuilder +putAllConfiguration(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringconfiguration) -ReplicationPeerConfigBuilder -setClusterKey(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringclusterKey) -Set the clusterKey which is the concatenation of the slave cluster's: - hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent - +default ReplicationPeerConfigBuilder +putAllPeerData(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Mapbyte[],byte[]peerData) ReplicationPeerConfigBuilder -setConfiguration(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringconfiguration) +putConfiguration(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringkey, +http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringvalue) ReplicationPeerConfigBuilder -setExcludeNamespaces(http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">Sethttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringnamespaces) +putPeerData(byte[]key, + byte[]value) ReplicationPeerConfigBuilder -setExcludeTableCFsMap(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapTableName,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringtableCFsMap) +setBandwidth(longbandwidth) ReplicationPeerConfigBuilder -setNamespaces(http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">Sethttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringnamespaces) +setClusterKey(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringclusterKey) +Set the clusterKey which is the concatenation of the slave cluster's: + hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent + ReplicationPeerConfigBuilder -setPeerData(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Mapbyte[],byte[]peerData) +setExcludeNamespaces(http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true;
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html index 6fecbc9..2accda0 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.CreateTableFuture.html @@ -34,4140 +34,4141 @@ 026import java.nio.charset.StandardCharsets; 027import java.util.ArrayList; 028import java.util.Arrays; -029import java.util.Collection; -030import java.util.EnumSet; -031import java.util.HashMap; -032import java.util.Iterator; -033import java.util.LinkedList; -034import java.util.List; -035import java.util.Map; -036import java.util.Set; -037import java.util.concurrent.Callable; -038import java.util.concurrent.ExecutionException; -039import java.util.concurrent.Future; -040import java.util.concurrent.TimeUnit; -041import java.util.concurrent.TimeoutException; -042import java.util.concurrent.atomic.AtomicInteger; -043import java.util.concurrent.atomic.AtomicReference; -044import java.util.regex.Pattern; -045import java.util.stream.Collectors; -046import java.util.stream.Stream; -047import org.apache.hadoop.conf.Configuration; -048import org.apache.hadoop.hbase.Abortable; -049import org.apache.hadoop.hbase.CacheEvictionStats; -050import org.apache.hadoop.hbase.CacheEvictionStatsBuilder; -051import org.apache.hadoop.hbase.ClusterMetrics.Option; -052import org.apache.hadoop.hbase.ClusterStatus; -053import org.apache.hadoop.hbase.DoNotRetryIOException; -054import org.apache.hadoop.hbase.HBaseConfiguration; -055import org.apache.hadoop.hbase.HConstants; -056import org.apache.hadoop.hbase.HRegionInfo; -057import org.apache.hadoop.hbase.HRegionLocation; -058import org.apache.hadoop.hbase.HTableDescriptor; -059import org.apache.hadoop.hbase.MasterNotRunningException; -060import org.apache.hadoop.hbase.MetaTableAccessor; -061import org.apache.hadoop.hbase.NamespaceDescriptor; -062import org.apache.hadoop.hbase.NamespaceNotFoundException; -063import org.apache.hadoop.hbase.NotServingRegionException; -064import org.apache.hadoop.hbase.RegionLoad; -065import org.apache.hadoop.hbase.RegionLocations; -066import org.apache.hadoop.hbase.ServerName; -067import org.apache.hadoop.hbase.TableExistsException; -068import org.apache.hadoop.hbase.TableName; -069import org.apache.hadoop.hbase.TableNotDisabledException; -070import org.apache.hadoop.hbase.TableNotFoundException; -071import org.apache.hadoop.hbase.UnknownRegionException; -072import org.apache.hadoop.hbase.ZooKeeperConnectionException; -073import org.apache.hadoop.hbase.client.replication.ReplicationPeerConfigUtil; -074import org.apache.hadoop.hbase.client.replication.TableCFs; -075import org.apache.hadoop.hbase.client.security.SecurityCapability; -076import org.apache.hadoop.hbase.exceptions.TimeoutIOException; -077import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel; -078import org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils; -079import org.apache.hadoop.hbase.ipc.HBaseRpcController; -080import org.apache.hadoop.hbase.ipc.RpcControllerFactory; -081import org.apache.hadoop.hbase.quotas.QuotaFilter; -082import org.apache.hadoop.hbase.quotas.QuotaRetriever; -083import org.apache.hadoop.hbase.quotas.QuotaSettings; -084import org.apache.hadoop.hbase.regionserver.wal.FailedLogCloseException; -085import org.apache.hadoop.hbase.replication.ReplicationException; -086import org.apache.hadoop.hbase.replication.ReplicationPeerConfig; -087import org.apache.hadoop.hbase.replication.ReplicationPeerDescription; -088import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils; -089import org.apache.hadoop.hbase.snapshot.HBaseSnapshotException; -090import org.apache.hadoop.hbase.snapshot.RestoreSnapshotException; -091import org.apache.hadoop.hbase.snapshot.SnapshotCreationException; -092import org.apache.hadoop.hbase.snapshot.UnknownSnapshotException; -093import org.apache.hadoop.hbase.util.Addressing; -094import org.apache.hadoop.hbase.util.Bytes; -095import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -096import org.apache.hadoop.hbase.util.ForeignExceptionUtil; -097import org.apache.hadoop.hbase.util.Pair; -098import org.apache.hadoop.ipc.RemoteException; -099import org.apache.hadoop.util.StringUtils; -100import org.apache.yetus.audience.InterfaceAudience; -101import org.apache.yetus.audience.InterfaceStability; -102import org.slf4j.Logger; -103import org.slf4j.LoggerFactory; -104 -105import org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting; -106import org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException; -107import
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4b2cc17/apidocs/src-html/org/apache/hadoop/hbase/client/Increment.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/client/Increment.html b/apidocs/src-html/org/apache/hadoop/hbase/client/Increment.html index 394abd6..ea801d2 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/client/Increment.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/client/Increment.html @@ -80,278 +80,290 @@ 072 } 073 /** 074 * Copy constructor -075 * @param i +075 * @param incrementToCopy increment to copy 076 */ -077 public Increment(Increment i) { -078this.row = i.getRow(); -079this.ts = i.getTimeStamp(); -080this.tr = i.getTimeRange(); -081 this.familyMap.putAll(i.getFamilyCellMap()); -082for (Map.EntryString, byte[] entry : i.getAttributesMap().entrySet()) { -083 this.setAttribute(entry.getKey(), entry.getValue()); -084} -085super.setPriority(i.getPriority()); -086 } -087 -088 /** -089 * Add the specified KeyValue to this operation. -090 * @param cell individual Cell -091 * @return this -092 * @throws java.io.IOException e -093 */ -094 public Increment add(Cell cell) throws IOException{ -095byte [] family = CellUtil.cloneFamily(cell); -096ListCell list = getCellList(family); -097//Checking that the row of the kv is the same as the put -098if (!CellUtil.matchingRows(cell, this.row)) { -099 throw new WrongRowIOException("The row in " + cell + -100" doesn't match the original one " + Bytes.toStringBinary(this.row)); -101} -102list.add(cell); -103return this; -104 } -105 -106 /** -107 * Increment the column from the specific family with the specified qualifier -108 * by the specified amount. -109 * p -110 * Overrides previous calls to addColumn for this family and qualifier. -111 * @param family family name -112 * @param qualifier column qualifier -113 * @param amount amount to increment by -114 * @return the Increment object -115 */ -116 public Increment addColumn(byte [] family, byte [] qualifier, long amount) { -117if (family == null) { -118 throw new IllegalArgumentException("family cannot be null"); -119} -120ListCell list = getCellList(family); -121KeyValue kv = createPutKeyValue(family, qualifier, ts, Bytes.toBytes(amount)); -122list.add(kv); -123return this; -124 } -125 -126 /** -127 * Gets the TimeRange used for this increment. -128 * @return TimeRange -129 */ -130 public TimeRange getTimeRange() { -131return this.tr; -132 } -133 -134 /** -135 * Sets the TimeRange to be used on the Get for this increment. -136 * p -137 * This is useful for when you have counters that only last for specific -138 * periods of time (ie. counters that are partitioned by time). By setting -139 * the range of valid times for this increment, you can potentially gain -140 * some performance with a more optimal Get operation. -141 * Be careful adding the time range to this class as you will update the old cell if the -142 * time range doesn't include the latest cells. -143 * p -144 * This range is used as [minStamp, maxStamp). -145 * @param minStamp minimum timestamp value, inclusive -146 * @param maxStamp maximum timestamp value, exclusive -147 * @throws IOException if invalid time range -148 * @return this -149 */ -150 public Increment setTimeRange(long minStamp, long maxStamp) -151 throws IOException { -152tr = new TimeRange(minStamp, maxStamp); -153return this; -154 } -155 -156 @Override -157 public Increment setTimestamp(long timestamp) { -158super.setTimestamp(timestamp); +077 public Increment(Increment incrementToCopy) { +078super(incrementToCopy); +079this.tr = incrementToCopy.getTimeRange(); +080 } +081 +082 /** +083 * Construct the Increment with user defined data. NOTED: +084 * 1) all cells in the familyMap must have the DataType.Put +085 * 2) the row of each cell must be same with passed row. +086 * @param row row. CAN'T be null +087 * @param ts timestamp +088 * @param familyMap the map to collect all cells internally. CAN'T be null +089 */ +090 public Increment(byte[] row, long ts, NavigableMapbyte [], ListCell familyMap) { +091super(row, ts, familyMap); +092 } +093 +094 /** +095 * Add the specified KeyValue to this operation. +096 * @param cell individual Cell +097 * @return this +098 * @throws java.io.IOException e +099 */ +100 public Increment add(Cell cell) throws IOException{ +101byte [] family = CellUtil.cloneFamily(cell); +102ListCell list = getCellList(family); +103//Checking that the row of the kv is the same as the put +104if (!CellUtil.matchingRows(cell, this.row)) { +105 throw new WrongRowIOException("The row in " + cell + +106" doesn't match the
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/505bbb2e/apidocs/src-html/org/apache/hadoop/hbase/util/Bytes.ByteArrayComparator.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/util/Bytes.ByteArrayComparator.html b/apidocs/src-html/org/apache/hadoop/hbase/util/Bytes.ByteArrayComparator.html index 51d92c2..86fc15e 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/util/Bytes.ByteArrayComparator.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/util/Bytes.ByteArrayComparator.html @@ -44,2578 +44,2580 @@ 036import java.util.Iterator; 037import java.util.List; 038 -039import com.google.protobuf.ByteString; -040import org.apache.commons.logging.Log; -041import org.apache.commons.logging.LogFactory; -042import org.apache.hadoop.hbase.Cell; -043import org.apache.hadoop.hbase.CellComparator; -044import org.apache.hadoop.hbase.KeyValue; -045import org.apache.hadoop.io.RawComparator; -046import org.apache.hadoop.io.WritableComparator; -047import org.apache.hadoop.io.WritableUtils; -048import org.apache.yetus.audience.InterfaceAudience; -049import sun.misc.Unsafe; -050 -051import org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting; -052import org.apache.hadoop.hbase.shaded.com.google.common.collect.Lists; +039import org.apache.hadoop.hbase.Cell; +040import org.apache.hadoop.hbase.CellComparator; +041import org.apache.hadoop.hbase.KeyValue; +042import org.apache.hadoop.io.RawComparator; +043import org.apache.hadoop.io.WritableComparator; +044import org.apache.hadoop.io.WritableUtils; +045import org.apache.yetus.audience.InterfaceAudience; +046import org.slf4j.Logger; +047import org.slf4j.LoggerFactory; +048 +049import org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting; +050import org.apache.hadoop.hbase.shaded.com.google.common.collect.Lists; +051 +052import com.google.protobuf.ByteString; 053 -054/** -055 * Utility class that handles byte arrays, conversions to/from other types, -056 * comparisons, hash code generation, manufacturing keys for HashMaps or -057 * HashSets, and can be used as key in maps or trees. -058 */ -059@SuppressWarnings("restriction") -060@InterfaceAudience.Public -061@edu.umd.cs.findbugs.annotations.SuppressWarnings( -062 value="EQ_CHECK_FOR_OPERAND_NOT_COMPATIBLE_WITH_THIS", -063justification="It has been like this forever") -064public class Bytes implements ComparableBytes { -065 -066 // Using the charset canonical name for String/byte[] conversions is much -067 // more efficient due to use of cached encoders/decoders. -068 private static final String UTF8_CSN = StandardCharsets.UTF_8.name(); -069 -070 //HConstants.EMPTY_BYTE_ARRAY should be updated if this changed -071 private static final byte [] EMPTY_BYTE_ARRAY = new byte [0]; -072 -073 private static final Log LOG = LogFactory.getLog(Bytes.class); +054import sun.misc.Unsafe; +055 +056/** +057 * Utility class that handles byte arrays, conversions to/from other types, +058 * comparisons, hash code generation, manufacturing keys for HashMaps or +059 * HashSets, and can be used as key in maps or trees. +060 */ +061@SuppressWarnings("restriction") +062@InterfaceAudience.Public +063@edu.umd.cs.findbugs.annotations.SuppressWarnings( +064 value="EQ_CHECK_FOR_OPERAND_NOT_COMPATIBLE_WITH_THIS", +065justification="It has been like this forever") +066public class Bytes implements ComparableBytes { +067 +068 // Using the charset canonical name for String/byte[] conversions is much +069 // more efficient due to use of cached encoders/decoders. +070 private static final String UTF8_CSN = StandardCharsets.UTF_8.name(); +071 +072 //HConstants.EMPTY_BYTE_ARRAY should be updated if this changed +073 private static final byte [] EMPTY_BYTE_ARRAY = new byte [0]; 074 -075 /** -076 * Size of boolean in bytes -077 */ -078 public static final int SIZEOF_BOOLEAN = Byte.SIZE / Byte.SIZE; -079 -080 /** -081 * Size of byte in bytes -082 */ -083 public static final int SIZEOF_BYTE = SIZEOF_BOOLEAN; -084 -085 /** -086 * Size of char in bytes -087 */ -088 public static final int SIZEOF_CHAR = Character.SIZE / Byte.SIZE; -089 -090 /** -091 * Size of double in bytes -092 */ -093 public static final int SIZEOF_DOUBLE = Double.SIZE / Byte.SIZE; -094 -095 /** -096 * Size of float in bytes -097 */ -098 public static final int SIZEOF_FLOAT = Float.SIZE / Byte.SIZE; -099 -100 /** -101 * Size of int in bytes -102 */ -103 public static final int SIZEOF_INT = Integer.SIZE / Byte.SIZE; -104 -105 /** -106 * Size of long in bytes -107 */ -108 public static final int SIZEOF_LONG = Long.SIZE / Byte.SIZE; -109 -110 /** -111 * Size of short in bytes -112 */ -113 public static final int SIZEOF_SHORT = Short.SIZE / Byte.SIZE; -114 -115 /** -116 * Mask to apply to a long to reveal the lower int only. Use like
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4c0cfa5/apidocs/src-html/org/apache/hadoop/hbase/filter/KeyOnlyFilter.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/filter/KeyOnlyFilter.html b/apidocs/src-html/org/apache/hadoop/hbase/filter/KeyOnlyFilter.html index 447ee2d..f943678 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/filter/KeyOnlyFilter.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/KeyOnlyFilter.html @@ -99,314 +99,316 @@ 091 /** 092 * @return The filter serialized using pb 093 */ -094 public byte [] toByteArray() { -095FilterProtos.KeyOnlyFilter.Builder builder = -096 FilterProtos.KeyOnlyFilter.newBuilder(); -097builder.setLenAsVal(this.lenAsVal); -098return builder.build().toByteArray(); -099 } -100 -101 /** -102 * @param pbBytes A pb serialized {@link KeyOnlyFilter} instance -103 * @return An instance of {@link KeyOnlyFilter} made from codebytes/code -104 * @throws DeserializationException -105 * @see #toByteArray -106 */ -107 public static KeyOnlyFilter parseFrom(final byte [] pbBytes) -108 throws DeserializationException { -109FilterProtos.KeyOnlyFilter proto; -110try { -111 proto = FilterProtos.KeyOnlyFilter.parseFrom(pbBytes); -112} catch (InvalidProtocolBufferException e) { -113 throw new DeserializationException(e); -114} -115return new KeyOnlyFilter(proto.getLenAsVal()); -116 } -117 -118 /** -119 * @param o the other filter to compare with -120 * @return true if and only if the fields of the filter that are serialized -121 * are equal to the corresponding fields in other. Used for testing. -122 */ -123 boolean areSerializedFieldsEqual(Filter o) { -124if (o == this) return true; -125if (!(o instanceof KeyOnlyFilter)) return false; -126 -127KeyOnlyFilter other = (KeyOnlyFilter)o; -128return this.lenAsVal == other.lenAsVal; -129 } -130 -131 static class KeyOnlyCell implements Cell { -132private Cell cell; -133private boolean lenAsVal; -134 -135public KeyOnlyCell(Cell c, boolean lenAsVal) { -136 this.cell = c; -137 this.lenAsVal = lenAsVal; -138} -139 -140@Override -141public byte[] getRowArray() { -142 return cell.getRowArray(); -143} -144 -145@Override -146public int getRowOffset() { -147 return cell.getRowOffset(); -148} -149 -150@Override -151public short getRowLength() { -152 return cell.getRowLength(); -153} -154 -155@Override -156public byte[] getFamilyArray() { -157 return cell.getFamilyArray(); -158} -159 -160@Override -161public int getFamilyOffset() { -162 return cell.getFamilyOffset(); -163} -164 -165@Override -166public byte getFamilyLength() { -167 return cell.getFamilyLength(); -168} -169 -170@Override -171public byte[] getQualifierArray() { -172 return cell.getQualifierArray(); -173} -174 -175@Override -176public int getQualifierOffset() { -177 return cell.getQualifierOffset(); -178} -179 -180@Override -181public int getQualifierLength() { -182 return cell.getQualifierLength(); -183} -184 -185@Override -186public long getTimestamp() { -187 return cell.getTimestamp(); -188} -189 -190@Override -191public byte getTypeByte() { -192 return cell.getTypeByte(); -193} -194 -195@Override -196public long getSequenceId() { -197 return 0; -198} -199 -200@Override -201public byte[] getValueArray() { -202 if (lenAsVal) { -203return Bytes.toBytes(cell.getValueLength()); -204 } else { -205return HConstants.EMPTY_BYTE_ARRAY; -206 } -207} -208 -209@Override -210public int getValueOffset() { -211 return 0; -212} -213 -214@Override -215public int getValueLength() { -216 if (lenAsVal) { -217return Bytes.SIZEOF_INT; -218 } else { -219return 0; -220 } -221} -222 -223@Override -224public byte[] getTagsArray() { -225 return HConstants.EMPTY_BYTE_ARRAY; -226} -227 -228@Override -229public int getTagsOffset() { -230 return 0; -231} -232 -233@Override -234public int getTagsLength() { -235 return 0; -236} -237 } -238 -239 static class KeyOnlyByteBufferCell extends ByteBufferCell { -240private ByteBufferCell cell; -241private boolean lenAsVal; -242 -243public KeyOnlyByteBufferCell(ByteBufferCell c, boolean lenAsVal) { -244 this.cell = c; -245 this.lenAsVal = lenAsVal; -246} -247 -248@Override -249public byte[] getRowArray() { -250 return cell.getRowArray(); -251} -252 -253@Override -254public int getRowOffset() { -255 return cell.getRowOffset(); -256} -257 -258@Override -259public short getRowLength() { -260 return
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/071f974b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html index f1a2443..a469e93 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html @@ -1350,415 +1350,415 @@ 1342return delete; 1343 } 1344 -1345 public static Put makeBarrierPut(byte[] encodedRegionName, long seq, byte[] tableName) { -1346byte[] seqBytes = Bytes.toBytes(seq); -1347return new Put(encodedRegionName) -1348 .addImmutable(HConstants.REPLICATION_BARRIER_FAMILY, seqBytes, seqBytes) -1349 .addImmutable(HConstants.REPLICATION_META_FAMILY, tableNameCq, tableName); -1350 } -1351 -1352 -1353 public static Put makeDaughterPut(byte[] encodedRegionName, byte[] value) { -1354return new Put(encodedRegionName).addImmutable(HConstants.REPLICATION_META_FAMILY, -1355daughterNameCq, value); -1356 } -1357 -1358 public static Put makeParentPut(byte[] encodedRegionName, byte[] value) { -1359return new Put(encodedRegionName).addImmutable(HConstants.REPLICATION_META_FAMILY, -1360parentNameCq, value); -1361 } -1362 -1363 /** -1364 * Adds split daughters to the Put -1365 */ -1366 public static Put addDaughtersToPut(Put put, RegionInfo splitA, RegionInfo splitB) { -1367if (splitA != null) { -1368 put.addImmutable( -1369HConstants.CATALOG_FAMILY, HConstants.SPLITA_QUALIFIER, RegionInfo.toByteArray(splitA)); -1370} -1371if (splitB != null) { -1372 put.addImmutable( -1373HConstants.CATALOG_FAMILY, HConstants.SPLITB_QUALIFIER, RegionInfo.toByteArray(splitB)); -1374} -1375return put; -1376 } -1377 -1378 /** -1379 * Put the passed codeputs/code to the codehbase:meta/code table. -1380 * Non-atomic for multi puts. -1381 * @param connection connection we're using -1382 * @param puts Put to add to hbase:meta -1383 * @throws IOException -1384 */ -1385 public static void putToMetaTable(final Connection connection, final Put... puts) -1386throws IOException { -1387put(getMetaHTable(connection), Arrays.asList(puts)); -1388 } -1389 -1390 /** -1391 * @param t Table to use (will be closed when done). -1392 * @param puts puts to make -1393 * @throws IOException -1394 */ -1395 private static void put(final Table t, final ListPut puts) throws IOException { -1396try { -1397 if (METALOG.isDebugEnabled()) { -1398 METALOG.debug(mutationsToString(puts)); -1399 } -1400 t.put(puts); -1401} finally { -1402 t.close(); -1403} -1404 } -1405 -1406 /** -1407 * Put the passed codeps/code to the codehbase:meta/code table. -1408 * @param connection connection we're using -1409 * @param ps Put to add to hbase:meta -1410 * @throws IOException -1411 */ -1412 public static void putsToMetaTable(final Connection connection, final ListPut ps) -1413throws IOException { -1414Table t = getMetaHTable(connection); -1415try { -1416 if (METALOG.isDebugEnabled()) { -1417 METALOG.debug(mutationsToString(ps)); -1418 } -1419 t.put(ps); -1420} finally { -1421 t.close(); -1422} -1423 } -1424 -1425 /** -1426 * Delete the passed coded/code from the codehbase:meta/code table. -1427 * @param connection connection we're using -1428 * @param d Delete to add to hbase:meta -1429 * @throws IOException -1430 */ -1431 static void deleteFromMetaTable(final Connection connection, final Delete d) -1432throws IOException { -1433ListDelete dels = new ArrayList(1); -1434dels.add(d); -1435deleteFromMetaTable(connection, dels); -1436 } -1437 -1438 /** -1439 * Delete the passed codedeletes/code from the codehbase:meta/code table. -1440 * @param connection connection we're using -1441 * @param deletes Deletes to add to hbase:meta This list should support #remove. -1442 * @throws IOException -1443 */ -1444 public static void deleteFromMetaTable(final Connection connection, final ListDelete deletes) -1445throws IOException { -1446Table t = getMetaHTable(connection); -1447try { -1448 if (METALOG.isDebugEnabled()) { -1449 METALOG.debug(mutationsToString(deletes)); -1450 } -1451 t.delete(deletes); -1452} finally { -1453 t.close(); -1454} -1455 } -1456 -1457 /** -1458 * Deletes some replica columns corresponding to replicas for the passed rows -1459 * @param metaRows rows in hbase:meta -1460 * @param replicaIndexToDeleteFrom the replica ID we would start deleting from -1461 * @param
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dc4e5c85/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.MergeTableRegionsFuture.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.MergeTableRegionsFuture.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.MergeTableRegionsFuture.html index 7c59e27..c904c56 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.MergeTableRegionsFuture.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.MergeTableRegionsFuture.html @@ -119,4048 +119,4054 @@ 111import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService; 112import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearCompactionQueuesRequest; 113import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ClearRegionBlockCacheRequest; -114import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest; -115import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest; -116import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest; -117import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse; -118import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest; -119import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse; -120import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest; -121import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest; -122import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos; -123import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest; -124import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse; -125import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos; -126import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription; -127import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType; -128import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema; -129import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos; -130import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest; -131import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse; -132import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest; -133import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse; -134import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest; -135import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest; -136import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest; -137import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse; -138import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest; -139import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse; -140import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest; -141import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse; -142import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest; -143import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse; -144import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest; -145import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest; -146import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse; -147import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest; -148import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse; -149import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest; -150import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse; -151import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest; -152import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse; -153import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest; -154import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest; -155import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest;
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4abd958d/devapidocs/org/apache/hadoop/hbase/client/HTable.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/HTable.html b/devapidocs/org/apache/hadoop/hbase/client/HTable.html index 380c1d7..a4d8801 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/HTable.html +++ b/devapidocs/org/apache/hadoop/hbase/client/HTable.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":42,"i8":10,"i9":42,"i10":10,"i11":10,"i12":42,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":9,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":9,"i33":10,"i34":10,"i35":10,"i36":9,"i37":10,"i38":42,"i39":10,"i40":10,"i41":42,"i42":10,"i43":10,"i44":42,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":42,"i51":42,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":42,"i61":42,"i62":42,"i63":42,"i64":10,"i65":10,"i66":9}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":42,"i7":42,"i8":42,"i9":10,"i10":42,"i11":42,"i12":42,"i13":42,"i14":42,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":9,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":9,"i34":10,"i35":10,"i36":10,"i37":9,"i38":10,"i39":42,"i40":10,"i41":10,"i42":42,"i43":10,"i44":10,"i45":42,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":42,"i52":42,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":42,"i62":42,"i63":42,"i64":42,"i65":10,"i66":10,"i67":9}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"],32:["t6","Deprecated Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab"; PrevClass -NextClass +NextClass Frames @@ -74,7 +74,7 @@ var activeTableTab = "activeTableTab"; Summary: -Nested| +Nested| Field| Constr| Method @@ -115,7 +115,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private @InterfaceStability.Stable -public class HTable +public class HTable extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements Table An implementation of Table. Used to communicate with a single HBase table. @@ -147,6 +147,32 @@ implements + + + + + +Nested Class Summary + +Nested Classes + +Modifier and Type +Class and Description + + +private class +HTable.CheckAndMutateBuilderImpl + + + + + + +Nested classes/interfaces inherited from interfaceorg.apache.hadoop.hbase.client.Table +Table.CheckAndMutateBuilder + + + @@ -336,8 +362,7 @@ implements Deletedelete) -Atomically checks if a row/family/qualifier value matches the expected - value. +Deprecated. @@ -359,11 +384,17 @@ implements CompareOperatorop, byte[]value, Deletedelete) -Atomically checks if a row/family/qualifier value matches the expected - value. +Deprecated. +Table.CheckAndMutateBuilder +checkAndMutate(byte[]row, + byte[]family) +Atomically checks if a row/family/qualifier value matches the expected value. + + + boolean checkAndMutate(byte[]row, byte[]family, @@ -374,7 +405,7 @@ implements Deprecated. - + boolean checkAndMutate(byte[]row, byte[]family, @@ -382,21 +413,20 @@ implements CompareOperatorop, byte[]value, RowMutationsrm) -Atomically checks if a row/family/qualifier value matches the expected value. +Deprecated. - + boolean checkAndPut(byte[]row, byte[]family, byte[]qualifier, byte[]value, Putput) -Atomically checks if a row/family/qualifier value matches the expected - value. +Deprecated. - + boolean checkAndPut(byte[]row, byte[]family, @@ -407,7 +437,7 @@ implements Deprecated. - + boolean checkAndPut(byte[]row, byte[]family, @@ -415,30 +445,29 @@ implements CompareOperatorop, byte[]value, Putput) -Atomically checks if a row/family/qualifier value matches the expected - value. +Deprecated. - + void clearRegionCache() Explicitly clears the region cache to fetch the latest value from META. - + void close() Releases any resources held or pending changes in internal buffers. - + CoprocessorRpcChannel coprocessorService(byte[]row) Creates and returns a RpcChannel instance connected to the table region containing the specified row. - + T extends com.google.protobuf.Service,Rhttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Mapbyte[],R
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e23b49ba/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html index a5af31c..a668610 100644 --- a/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html +++ b/devapidocs/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -119,7 +119,7 @@ var activeTableTab = "activeTableTab"; All Implemented Interfaces: -Cell, SettableSequenceId +http://docs.oracle.com/javase/8/docs/api/java/lang/Cloneable.html?is-external=true; title="class or interface in java.lang">Cloneable, Cell, ExtendedCell, HeapSize, RawCell Direct Known Subclasses: @@ -131,7 +131,7 @@ var activeTableTab = "activeTableTab"; -private static class PrivateCellUtil.FirstOnRowColCell +private static class PrivateCellUtil.FirstOnRowColCell extends PrivateCellUtil.FirstOnRowCell @@ -156,26 +156,44 @@ extends fArray +private static long +FIXED_HEAPSIZE + + private byte flength - + private int foffset - + private byte[] qArray - + private int qlength - + private int qoffset + + + + +Fields inherited from interfaceorg.apache.hadoop.hbase.ExtendedCell +CELL_NOT_BASED_ON_CHUNK + + + + + +Fields inherited from interfaceorg.apache.hadoop.hbase.RawCell +MAX_TAGS_LENGTH + @@ -244,6 +262,10 @@ extends int getQualifierOffset() + +long +heapSize() + @@ -257,7 +279,7 @@ extends PrivateCellUtil.EmptyCell -getSequenceId, getTagsArray, getTagsLength, getTagsOffset, getValueArray, getValueLength, getValueOffset, setSequenceId +getSequenceId, getTagsArray, getTagsLength, getTagsOffset, getValueArray, getValueLength, getValueOffset, setSequenceId, setTimestamp, setTimestamp @@ -266,6 +288,20 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--; title="class or interface in java.lang">clone, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-; title="class or interface in java.lang">equals, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--; title="class or interface in java.lang">finalize, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--; title="class or interface in java.lang">getClass, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--; title="class or interface in java.lang">hashCode, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--; title="class or interface in java.lang">notify, http://docs.oracle.com/javase/8/docs/api/java/lang /Object.html?is-external=true#notifyAll--" title="class or interface in java.lang">notifyAll, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--; title="class or interface in java.lang">toString, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait--; title="class or interface in java.lang">wait, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-; title="class or interface in java.lang">wait, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-int-; title="class or interface in java.lang">wait + + + + +Methods inherited from interfaceorg.apache.hadoop.hbase.ExtendedCell +deepClone, getChunkId, getSerializedSize, write, write + + + + + +Methods inherited from interfaceorg.apache.hadoop.hbase.RawCell +checkForTagsLength, cloneTags, getTag, getTags + @@ -280,13 +316,22 @@ extends + + + + +FIXED_HEAPSIZE +private static finallong FIXED_HEAPSIZE + + fArray -private finalbyte[] fArray +private finalbyte[] fArray @@ -295,7 +340,7 @@ extends foffset -private finalint foffset +private finalint foffset @@ -304,7 +349,7 @@ extends flength -private finalbyte flength +private finalbyte flength @@ -313,7 +358,7 @@ extends qArray -private finalbyte[] qArray +private finalbyte[] qArray @@ -322,7 +367,7 @@ extends qoffset -private finalint qoffset +private finalint qoffset @@ -331,7 +376,7 @@ extends qlength -private finalint qlength +private finalint qlength @@ -348,7 +393,7
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/RegionObserver.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/RegionObserver.html b/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/RegionObserver.html index 4c37cbe..cfb8ee4 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/RegionObserver.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/RegionObserver.html @@ -228,837 +228,853 @@ 220 * of candidates. If you remove all the candidates then the compaction will be canceled. 221 * pSupports Coprocessor 'bypass' -- 'bypass' is how this method indicates that it changed 222 * the passed in codecandidates/code. -223 * @param c the environment provided by the region server -224 * @param store the store where compaction is being requested -225 * @param candidates the store files currently available for compaction -226 * @param tracker tracker used to track the life cycle of a compaction -227 */ -228 default void preCompactSelection(ObserverContextRegionCoprocessorEnvironment c, Store store, -229 List? extends StoreFile candidates, CompactionLifeCycleTracker tracker) -230 throws IOException {} -231 -232 /** -233 * Called after the {@link StoreFile}s to compact have been selected from the available -234 * candidates. -235 * @param c the environment provided by the region server -236 * @param store the store being compacted -237 * @param selected the store files selected to compact -238 * @param tracker tracker used to track the life cycle of a compaction -239 * @param request the requested compaction -240 */ -241 default void postCompactSelection(ObserverContextRegionCoprocessorEnvironment c, Store store, -242 List? extends StoreFile selected, CompactionLifeCycleTracker tracker, -243 CompactionRequest request) {} -244 -245 /** -246 * Called before we open store scanner for compaction. You can use the {@code options} to change max -247 * versions and TTL for the scanner being opened. -248 * @param c the environment provided by the region server -249 * @param store the store being compacted -250 * @param scanType type of Scan -251 * @param options used to change max versions and TTL for the scanner being opened -252 * @param tracker tracker used to track the life cycle of a compaction -253 * @param request the requested compaction -254 */ -255 default void preCompactScannerOpen(ObserverContextRegionCoprocessorEnvironment c, Store store, -256 ScanType scanType, ScanOptions options, CompactionLifeCycleTracker tracker, -257 CompactionRequest request) throws IOException {} -258 -259 /** -260 * Called prior to writing the {@link StoreFile}s selected for compaction into a new -261 * {@code StoreFile}. -262 * p -263 * To override or modify the compaction process, implementing classes can wrap the provided -264 * {@link InternalScanner} with a custom implementation that is returned from this method. The -265 * custom scanner can then inspect {@link org.apache.hadoop.hbase.Cell}s from the wrapped scanner, -266 * applying its own policy to what gets written. -267 * @param c the environment provided by the region server -268 * @param store the store being compacted -269 * @param scanner the scanner over existing data used in the store file rewriting -270 * @param scanType type of Scan -271 * @param tracker tracker used to track the life cycle of a compaction -272 * @param request the requested compaction -273 * @return the scanner to use during compaction. Should not be {@code null} unless the -274 * implementation is writing new store files on its own. -275 */ -276 default InternalScanner preCompact(ObserverContextRegionCoprocessorEnvironment c, Store store, -277 InternalScanner scanner, ScanType scanType, CompactionLifeCycleTracker tracker, -278 CompactionRequest request) throws IOException { -279return scanner; -280 } -281 -282 /** -283 * Called after compaction has completed and the new store file has been moved in to place. -284 * @param c the environment provided by the region server -285 * @param store the store being compacted -286 * @param resultFile the new store file written out during compaction -287 * @param tracker used to track the life cycle of a compaction -288 * @param request the requested compaction -289 */ -290 default void postCompact(ObserverContextRegionCoprocessorEnvironment c, Store store, -291 StoreFile resultFile, CompactionLifeCycleTracker tracker, CompactionRequest request) -292 throws IOException {} -293 -294 /** -295 * Called before the region is reported as closed to the master. -296 * @param c the environment provided by the region server -297 * @param
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c54c242b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html index 3edfbef..9707b2c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.PrepareFlushResult.html @@ -2459,5936 +2459,5935 @@ 2451 } 2452 2453 for (HStore s : storesToFlush) { -2454MemStoreSize flushableSize = s.getFlushableSize(); -2455 totalSizeOfFlushableStores.incMemStoreSize(flushableSize); -2456 storeFlushCtxs.put(s.getColumnFamilyDescriptor().getName(), -2457 s.createFlushContext(flushOpSeqId, tracker)); -2458// for writing stores to WAL -2459 committedFiles.put(s.getColumnFamilyDescriptor().getName(), null); -2460 storeFlushableSize.put(s.getColumnFamilyDescriptor().getName(), flushableSize); -2461 } -2462 -2463 // write the snapshot start to WAL -2464 if (wal != null !writestate.readOnly) { -2465FlushDescriptor desc = ProtobufUtil.toFlushDescriptor(FlushAction.START_FLUSH, -2466getRegionInfo(), flushOpSeqId, committedFiles); -2467// No sync. Sync is below where no updates lock and we do FlushAction.COMMIT_FLUSH -2468WALUtil.writeFlushMarker(wal, this.getReplicationScope(), getRegionInfo(), desc, false, -2469mvcc); -2470 } -2471 -2472 // Prepare flush (take a snapshot) -2473 for (StoreFlushContext flush : storeFlushCtxs.values()) { -2474flush.prepare(); -2475 } -2476} catch (IOException ex) { -2477 doAbortFlushToWAL(wal, flushOpSeqId, committedFiles); -2478 throw ex; -2479} finally { -2480 this.updatesLock.writeLock().unlock(); -2481} -2482String s = "Finished memstore snapshotting " + this + ", syncing WAL and waiting on mvcc, " + -2483"flushsize=" + totalSizeOfFlushableStores; -2484status.setStatus(s); -2485doSyncOfUnflushedWALChanges(wal, getRegionInfo()); -2486return new PrepareFlushResult(storeFlushCtxs, committedFiles, storeFlushableSize, startTime, -2487flushOpSeqId, flushedSeqId, totalSizeOfFlushableStores); -2488 } -2489 -2490 /** -2491 * Utility method broken out of internalPrepareFlushCache so that method is smaller. -2492 */ -2493 private void logFatLineOnFlush(CollectionHStore storesToFlush, long sequenceId) { -2494if (!LOG.isInfoEnabled()) { -2495 return; -2496} -2497// Log a fat line detailing what is being flushed. -2498StringBuilder perCfExtras = null; -2499if (!isAllFamilies(storesToFlush)) { -2500 perCfExtras = new StringBuilder(); -2501 for (HStore store: storesToFlush) { -2502perCfExtras.append("; ").append(store.getColumnFamilyName()); -2503perCfExtras.append("=") -2504 .append(StringUtils.byteDesc(store.getFlushableSize().getDataSize())); -2505 } -2506} -2507LOG.info("Flushing " + + storesToFlush.size() + "/" + stores.size() + -2508" column families, memstore=" + StringUtils.byteDesc(this.memstoreDataSize.get()) + -2509((perCfExtras != null perCfExtras.length() 0)? perCfExtras.toString(): "") + -2510((wal != null) ? "" : "; WAL is null, using passed sequenceid=" + sequenceId)); -2511 } -2512 -2513 private void doAbortFlushToWAL(final WAL wal, final long flushOpSeqId, -2514 final Mapbyte[], ListPath committedFiles) { -2515if (wal == null) return; -2516try { -2517 FlushDescriptor desc = ProtobufUtil.toFlushDescriptor(FlushAction.ABORT_FLUSH, -2518 getRegionInfo(), flushOpSeqId, committedFiles); -2519 WALUtil.writeFlushMarker(wal, this.getReplicationScope(), getRegionInfo(), desc, false, -2520 mvcc); -2521} catch (Throwable t) { -2522 LOG.warn("Received unexpected exception trying to write ABORT_FLUSH marker to WAL:" + -2523 StringUtils.stringifyException(t)); -2524 // ignore this since we will be aborting the RS with DSE. -2525} -2526// we have called wal.startCacheFlush(), now we have to abort it -2527 wal.abortCacheFlush(this.getRegionInfo().getEncodedNameAsBytes()); -2528 } -2529 -2530 /** -2531 * Sync unflushed WAL changes. See HBASE-8208 for details -2532 */ -2533 private static void doSyncOfUnflushedWALChanges(final WAL wal, final RegionInfo hri) -2534 throws IOException { -2535if (wal == null) { -2536 return; -2537} -2538try { -2539 wal.sync(); // ensure that flush marker is sync'ed -2540} catch (IOException ioe) { -2541
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713d773f/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.html b/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.html index 11bfb15..915e78a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/HConstants.html @@ -782,557 +782,562 @@ 774 /** 775 * Default value of {@link #HBASE_CLIENT_RETRIES_NUMBER}. 776 */ -777 public static final int DEFAULT_HBASE_CLIENT_RETRIES_NUMBER = 35; +777 public static final int DEFAULT_HBASE_CLIENT_RETRIES_NUMBER = 10; 778 -779 /** -780 * Parameter name to set the default scanner caching for all clients. -781 */ -782 public static final String HBASE_CLIENT_SCANNER_CACHING = "hbase.client.scanner.caching"; +779 public static final String HBASE_CLIENT_SERVERSIDE_RETRIES_MULTIPLIER = +780 "hbase.client.serverside.retries.multiplier"; +781 +782 public static final int DEFAULT_HBASE_CLIENT_SERVERSIDE_RETRIES_MULTIPLIER = 3; 783 784 /** -785 * Default value for {@link #HBASE_CLIENT_SCANNER_CACHING} +785 * Parameter name to set the default scanner caching for all clients. 786 */ -787 public static final int DEFAULT_HBASE_CLIENT_SCANNER_CACHING = Integer.MAX_VALUE; +787 public static final String HBASE_CLIENT_SCANNER_CACHING = "hbase.client.scanner.caching"; 788 789 /** -790 * Parameter name for number of rows that will be fetched when calling next on -791 * a scanner if it is not served from memory. Higher caching values will -792 * enable faster scanners but will eat up more memory and some calls of next -793 * may take longer and longer times when the cache is empty. -794 */ -795 public static final String HBASE_META_SCANNER_CACHING = "hbase.meta.scanner.caching"; -796 -797 /** -798 * Default value of {@link #HBASE_META_SCANNER_CACHING}. +790 * Default value for {@link #HBASE_CLIENT_SCANNER_CACHING} +791 */ +792 public static final int DEFAULT_HBASE_CLIENT_SCANNER_CACHING = Integer.MAX_VALUE; +793 +794 /** +795 * Parameter name for number of rows that will be fetched when calling next on +796 * a scanner if it is not served from memory. Higher caching values will +797 * enable faster scanners but will eat up more memory and some calls of next +798 * may take longer and longer times when the cache is empty. 799 */ -800 public static final int DEFAULT_HBASE_META_SCANNER_CACHING = 100; +800 public static final String HBASE_META_SCANNER_CACHING = "hbase.meta.scanner.caching"; 801 802 /** -803 * Parameter name for number of versions, kept by meta table. +803 * Default value of {@link #HBASE_META_SCANNER_CACHING}. 804 */ -805 public static final String HBASE_META_VERSIONS = "hbase.meta.versions"; +805 public static final int DEFAULT_HBASE_META_SCANNER_CACHING = 100; 806 807 /** -808 * Default value of {@link #HBASE_META_VERSIONS}. +808 * Parameter name for number of versions, kept by meta table. 809 */ -810 public static final int DEFAULT_HBASE_META_VERSIONS = 3; +810 public static final String HBASE_META_VERSIONS = "hbase.meta.versions"; 811 812 /** -813 * Parameter name for number of versions, kept by meta table. +813 * Default value of {@link #HBASE_META_VERSIONS}. 814 */ -815 public static final String HBASE_META_BLOCK_SIZE = "hbase.meta.blocksize"; +815 public static final int DEFAULT_HBASE_META_VERSIONS = 3; 816 817 /** -818 * Default value of {@link #HBASE_META_BLOCK_SIZE}. +818 * Parameter name for number of versions, kept by meta table. 819 */ -820 public static final int DEFAULT_HBASE_META_BLOCK_SIZE = 8 * 1024; +820 public static final String HBASE_META_BLOCK_SIZE = "hbase.meta.blocksize"; 821 822 /** -823 * Parameter name for unique identifier for this {@link org.apache.hadoop.conf.Configuration} -824 * instance. If there are two or more {@link org.apache.hadoop.conf.Configuration} instances that, -825 * for all intents and purposes, are the same except for their instance ids, then they will not be -826 * able to share the same org.apache.hadoop.hbase.client.HConnection instance. On the other hand, -827 * even if the instance ids are the same, it could result in non-shared -828 * org.apache.hadoop.hbase.client.HConnection instances if some of the other connection parameters -829 * differ. -830 */ -831 public static final String HBASE_CLIENT_INSTANCE_ID = "hbase.client.instance.id"; -832 -833 /** -834 * The client scanner timeout period in milliseconds. +823 * Default value of {@link #HBASE_META_BLOCK_SIZE}. +824 */ +825 public static final int DEFAULT_HBASE_META_BLOCK_SIZE = 8 * 1024; +826 +827 /** +828 * Parameter name for unique identifier for this {@link org.apache.hadoop.conf.Configuration}
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fd365a2b/devapidocs/org/apache/hadoop/hbase/thrift/TBoundedThreadPoolServer.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/thrift/TBoundedThreadPoolServer.html b/devapidocs/org/apache/hadoop/hbase/thrift/TBoundedThreadPoolServer.html index 741fc75..a0a7ebb 100644 --- a/devapidocs/org/apache/hadoop/hbase/thrift/TBoundedThreadPoolServer.html +++ b/devapidocs/org/apache/hadoop/hbase/thrift/TBoundedThreadPoolServer.html @@ -115,7 +115,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public class TBoundedThreadPoolServer +public class TBoundedThreadPoolServer extends org.apache.thrift.server.TServer A bounded thread pool server customized for HBase. @@ -332,7 +332,7 @@ extends org.apache.thrift.server.TServer QUEUE_FULL_MSG -private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String QUEUE_FULL_MSG +private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String QUEUE_FULL_MSG See Also: Constant Field Values @@ -345,7 +345,7 @@ extends org.apache.thrift.server.TServer MIN_WORKER_THREADS_CONF_KEY -public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String MIN_WORKER_THREADS_CONF_KEY +public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String MIN_WORKER_THREADS_CONF_KEY The "core size" of the thread pool. New threads are created on every connection until this many threads are created. @@ -360,7 +360,7 @@ extends org.apache.thrift.server.TServer DEFAULT_MIN_WORKER_THREADS -public static finalint DEFAULT_MIN_WORKER_THREADS +public static finalint DEFAULT_MIN_WORKER_THREADS This default core pool size should be enough for many test scenarios. We want to override this with a much larger number (e.g. at least 200) for a large-scale production setup. @@ -376,7 +376,7 @@ extends org.apache.thrift.server.TServer MAX_WORKER_THREADS_CONF_KEY -public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String MAX_WORKER_THREADS_CONF_KEY +public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String MAX_WORKER_THREADS_CONF_KEY The maximum size of the thread pool. When the pending request queue overflows, new threads are created until their number reaches this number. After that, the server starts dropping connections. @@ -392,7 +392,7 @@ extends org.apache.thrift.server.TServer DEFAULT_MAX_WORKER_THREADS -public static finalint DEFAULT_MAX_WORKER_THREADS +public static finalint DEFAULT_MAX_WORKER_THREADS See Also: Constant Field Values @@ -405,7 +405,7 @@ extends org.apache.thrift.server.TServer MAX_QUEUED_REQUESTS_CONF_KEY -public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String MAX_QUEUED_REQUESTS_CONF_KEY +public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String MAX_QUEUED_REQUESTS_CONF_KEY The maximum number of pending connections waiting in the queue. If there are no idle threads in the pool, the server queues requests. Only when the queue overflows, new threads are added, up to @@ -422,7 +422,7 @@ extends org.apache.thrift.server.TServer DEFAULT_MAX_QUEUED_REQUESTS -public static finalint DEFAULT_MAX_QUEUED_REQUESTS +public static finalint DEFAULT_MAX_QUEUED_REQUESTS See Also: Constant Field Values @@ -435,7 +435,7 @@ extends org.apache.thrift.server.TServer THREAD_KEEP_ALIVE_TIME_SEC_CONF_KEY -public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String THREAD_KEEP_ALIVE_TIME_SEC_CONF_KEY +public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String THREAD_KEEP_ALIVE_TIME_SEC_CONF_KEY Default amount of time in seconds to keep a thread alive. Worker threads are stopped after being idle for this long. @@ -450,7 +450,7 @@ extends org.apache.thrift.server.TServer DEFAULT_THREAD_KEEP_ALIVE_TIME_SEC -private static finalint DEFAULT_THREAD_KEEP_ALIVE_TIME_SEC +private static finalint DEFAULT_THREAD_KEEP_ALIVE_TIME_SEC See Also: Constant Field Values @@ -463,7 +463,7 @@ extends org.apache.thrift.server.TServer TIME_TO_WAIT_AFTER_SHUTDOWN_MS -public static finalint TIME_TO_WAIT_AFTER_SHUTDOWN_MS +public
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b9722a17/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html index 40cd159..2da0903 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionStat.html @@ -260,7 +260,7 @@ 252 253// Update meta events (for testing) 254if (hasProcExecutor) { -255 getProcedureScheduler().suspendEvent(metaLoadEvent); +255 metaLoadEvent.suspend(); 256 setFailoverCleanupDone(false); 257 for (RegionInfo hri: getMetaRegionSet()) { 258setMetaInitialized(hri, false); @@ -421,1455 +421,1454 @@ 413 } 414 415 public boolean waitMetaInitialized(final Procedure proc, final RegionInfo regionInfo) { -416return getProcedureScheduler().waitEvent( -417 getMetaInitializedEvent(getMetaForRegion(regionInfo)), proc); -418 } -419 -420 private void setMetaInitialized(final RegionInfo metaRegionInfo, final boolean isInitialized) { -421assert isMetaRegion(metaRegionInfo) : "unexpected non-meta region " + metaRegionInfo; -422final ProcedureEvent metaInitEvent = getMetaInitializedEvent(metaRegionInfo); -423if (isInitialized) { -424 getProcedureScheduler().wakeEvent(metaInitEvent); -425} else { -426 getProcedureScheduler().suspendEvent(metaInitEvent); -427} -428 } -429 -430 private ProcedureEvent getMetaInitializedEvent(final RegionInfo metaRegionInfo) { -431assert isMetaRegion(metaRegionInfo) : "unexpected non-meta region " + metaRegionInfo; -432// TODO: handle multiple meta. -433return metaInitializedEvent; -434 } -435 -436 public boolean waitMetaLoaded(final Procedure proc) { -437return getProcedureScheduler().waitEvent(metaLoadEvent, proc); -438 } -439 -440 protected void wakeMetaLoadedEvent() { -441 getProcedureScheduler().wakeEvent(metaLoadEvent); -442assert isMetaLoaded() : "expected meta to be loaded"; -443 } -444 -445 public boolean isMetaLoaded() { -446return metaLoadEvent.isReady(); -447 } -448 -449 // -450 // TODO: Sync helpers -451 // -452 public void assignMeta(final RegionInfo metaRegionInfo) throws IOException { -453assignMeta(metaRegionInfo, null); -454 } -455 -456 public void assignMeta(final RegionInfo metaRegionInfo, final ServerName serverName) -457 throws IOException { -458assert isMetaRegion(metaRegionInfo) : "unexpected non-meta region " + metaRegionInfo; -459AssignProcedure proc; -460if (serverName != null) { -461 LOG.debug("Try assigning Meta " + metaRegionInfo + " to " + serverName); -462 proc = createAssignProcedure(metaRegionInfo, serverName); -463} else { -464 LOG.debug("Assigning " + metaRegionInfo.getRegionNameAsString()); -465 proc = createAssignProcedure(metaRegionInfo, false); -466} -467 ProcedureSyncWait.submitAndWaitProcedure(master.getMasterProcedureExecutor(), proc); -468 } -469 -470 /** -471 * Start a new thread to check if there are region servers whose versions are higher than others. -472 * If so, move all system table regions to RS with the highest version to keep compatibility. -473 * The reason is, RS in new version may not be able to access RS in old version when there are -474 * some incompatible changes. -475 */ -476 public void checkIfShouldMoveSystemRegionAsync() { -477new Thread(() - { -478 try { -479synchronized (checkIfShouldMoveSystemRegionLock) { -480 ListRegionPlan plans = new ArrayList(); -481 for (ServerName server : getExcludedServersForSystemTable()) { -482if (master.getServerManager().isServerDead(server)) { -483 // TODO: See HBASE-18494 and HBASE-18495. Though getExcludedServersForSystemTable() -484 // considers only online servers, the server could be queued for dead server -485 // processing. As region assignments for crashed server is handled by -486 // ServerCrashProcedure, do NOT handle them here. The goal is to handle this through -487 // regular flow of LoadBalancer as a favored node and not to have this special -488 // handling. -489 continue; -490} -491ListRegionInfo regionsShouldMove = getCarryingSystemTables(server); -492
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b1eb7453/devapidocs/org/apache/hadoop/hbase/regionserver/HMobStore.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HMobStore.html b/devapidocs/org/apache/hadoop/hbase/regionserver/HMobStore.html index d941aa1..7b68289 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HMobStore.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HMobStore.html @@ -119,7 +119,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public class HMobStore +public class HMobStore extends HStore The store implementation to save MOBs (medium objects), it extends the HStore. When a descriptor of a column family has the value "IS_MOB", it means this column family @@ -501,7 +501,7 @@ extends LOG -private static finalorg.apache.commons.logging.Log LOG +private static finalorg.apache.commons.logging.Log LOG @@ -510,7 +510,7 @@ extends mobCacheConfig -privateMobCacheConfig mobCacheConfig +privateMobCacheConfig mobCacheConfig @@ -519,7 +519,7 @@ extends homePath -privateorg.apache.hadoop.fs.Path homePath +privateorg.apache.hadoop.fs.Path homePath @@ -528,7 +528,7 @@ extends mobFamilyPath -privateorg.apache.hadoop.fs.Path mobFamilyPath +privateorg.apache.hadoop.fs.Path mobFamilyPath @@ -537,7 +537,7 @@ extends cellsCountCompactedToMob -private volatilelong cellsCountCompactedToMob +private volatilelong cellsCountCompactedToMob @@ -546,7 +546,7 @@ extends cellsCountCompactedFromMob -private volatilelong cellsCountCompactedFromMob +private volatilelong cellsCountCompactedFromMob @@ -555,7 +555,7 @@ extends cellsSizeCompactedToMob -private volatilelong cellsSizeCompactedToMob +private volatilelong cellsSizeCompactedToMob @@ -564,7 +564,7 @@ extends cellsSizeCompactedFromMob -private volatilelong cellsSizeCompactedFromMob +private volatilelong cellsSizeCompactedFromMob @@ -573,7 +573,7 @@ extends mobFlushCount -private volatilelong mobFlushCount +private volatilelong mobFlushCount @@ -582,7 +582,7 @@ extends mobFlushedCellsCount -private volatilelong mobFlushedCellsCount +private volatilelong mobFlushedCellsCount @@ -591,7 +591,7 @@ extends mobFlushedCellsSize -private volatilelong mobFlushedCellsSize +private volatilelong mobFlushedCellsSize @@ -600,7 +600,7 @@ extends mobScanCellsCount -private volatilelong mobScanCellsCount +private volatilelong mobScanCellsCount @@ -609,7 +609,7 @@ extends mobScanCellsSize -private volatilelong mobScanCellsSize +private volatilelong mobScanCellsSize @@ -618,7 +618,7 @@ extends family -privateColumnFamilyDescriptor family +privateColumnFamilyDescriptor family @@ -627,7 +627,7 @@ extends map -privatehttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.fs.Path map +privatehttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.fs.Path map @@ -636,7 +636,7 @@ extends keyLock -private finalIdLock keyLock +private finalIdLock keyLock @@ -645,7 +645,7 @@ extends refCellTags -private finalbyte[] refCellTags +private finalbyte[] refCellTags @@ -662,7 +662,7 @@ extends HMobStore -publicHMobStore(HRegionregion, +publicHMobStore(HRegionregion, ColumnFamilyDescriptorfamily, org.apache.hadoop.conf.ConfigurationconfParam) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException @@ -686,7 +686,7 @@ extends createCacheConf -protectedvoidcreateCacheConf(ColumnFamilyDescriptorfamily) +protectedvoidcreateCacheConf(ColumnFamilyDescriptorfamily) Creates the mob cache config. Overrides: @@ -702,7 +702,7 @@ extends getConfiguration -publicorg.apache.hadoop.conf.ConfigurationgetConfiguration() +publicorg.apache.hadoop.conf.ConfigurationgetConfiguration() Gets current config. @@ -712,7 +712,7 @@ extends createScanner -protectedKeyValueScannercreateScanner(Scanscan, +protectedKeyValueScannercreateScanner(Scanscan, ScanInfoscanInfo,
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/9118853f/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionObserverOperationWithoutResult.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionObserverOperationWithoutResult.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionObserverOperationWithoutResult.html index a9e8ff7..c670860 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionObserverOperationWithoutResult.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.RegionObserverOperationWithoutResult.html @@ -65,1723 +65,1733 @@ 057import org.apache.hadoop.hbase.client.TableDescriptor; 058import org.apache.hadoop.hbase.coprocessor.BaseEnvironment; 059import org.apache.hadoop.hbase.coprocessor.BulkLoadObserver; -060import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; -061import org.apache.hadoop.hbase.coprocessor.CoprocessorService; -062import org.apache.hadoop.hbase.coprocessor.CoprocessorServiceBackwardCompatiblity; -063import org.apache.hadoop.hbase.coprocessor.CoreCoprocessor; -064import org.apache.hadoop.hbase.coprocessor.EndpointObserver; -065import org.apache.hadoop.hbase.coprocessor.HasRegionServerServices; -066import org.apache.hadoop.hbase.coprocessor.MetricsCoprocessor; -067import org.apache.hadoop.hbase.coprocessor.ObserverContext; -068import org.apache.hadoop.hbase.coprocessor.RegionCoprocessor; -069import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment; -070import org.apache.hadoop.hbase.coprocessor.RegionObserver; -071import org.apache.hadoop.hbase.coprocessor.RegionObserver.MutationType; -072import org.apache.hadoop.hbase.filter.ByteArrayComparable; -073import org.apache.hadoop.hbase.io.FSDataInputStreamWrapper; -074import org.apache.hadoop.hbase.io.Reference; -075import org.apache.hadoop.hbase.io.hfile.CacheConfig; -076import org.apache.hadoop.hbase.metrics.MetricRegistry; -077import org.apache.hadoop.hbase.regionserver.Region.Operation; -078import org.apache.hadoop.hbase.regionserver.compactions.CompactionLifeCycleTracker; -079import org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest; -080import org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker; -081import org.apache.hadoop.hbase.security.User; -082import org.apache.hadoop.hbase.util.Bytes; -083import org.apache.hadoop.hbase.util.CoprocessorClassLoader; -084import org.apache.hadoop.hbase.util.Pair; -085import org.apache.hadoop.hbase.wal.WALEdit; -086import org.apache.hadoop.hbase.wal.WALKey; -087import org.apache.yetus.audience.InterfaceAudience; -088 -089import org.apache.hadoop.hbase.shaded.com.google.common.collect.Lists; -090 -091/** -092 * Implements the coprocessor environment and runtime support for coprocessors -093 * loaded within a {@link Region}. -094 */ -095@InterfaceAudience.Private -096public class RegionCoprocessorHost -097extends CoprocessorHostRegionCoprocessor, RegionCoprocessorEnvironment { -098 -099 private static final Log LOG = LogFactory.getLog(RegionCoprocessorHost.class); -100 // The shared data map -101 private static final ReferenceMapString, ConcurrentMapString, Object SHARED_DATA_MAP = -102 new ReferenceMap(AbstractReferenceMap.ReferenceStrength.HARD, -103 AbstractReferenceMap.ReferenceStrength.WEAK); -104 -105 // optimization: no need to call postScannerFilterRow, if no coprocessor implements it -106 private final boolean hasCustomPostScannerFilterRow; -107 -108 /** -109 * -110 * Encapsulation of the environment of each coprocessor -111 */ -112 private static class RegionEnvironment extends BaseEnvironmentRegionCoprocessor -113 implements RegionCoprocessorEnvironment { -114private Region region; -115ConcurrentMapString, Object sharedData; -116private final MetricRegistry metricRegistry; -117private final RegionServerServices services; -118 -119/** -120 * Constructor -121 * @param impl the coprocessor instance -122 * @param priority chaining priority -123 */ -124public RegionEnvironment(final RegionCoprocessor impl, final int priority, -125final int seq, final Configuration conf, final Region region, -126final RegionServerServices services, final ConcurrentMapString, Object sharedData) { -127 super(impl, priority, seq, conf); -128 this.region = region; -129 this.sharedData = sharedData; -130 this.services = services; -131 this.metricRegistry = -132 MetricsCoprocessor.createRegistryForRegionCoprocessor(impl.getClass().getName()); -133} -134 -135/** @return the region */ -136@Override -137public Region getRegion() { -138 return region; -139} -140 -141public
1 2 3 >