[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/346adc37/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html index b3d18d1..1c9f960 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html @@ -196,7 +196,7 @@ UnassignProcedure -Procedure that describe the unassignment of a single region. +Procedure that describes the unassignment of a single region. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/346adc37/devapidocs/org/apache/hadoop/hbase/master/assignment/package-use.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-use.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-use.html index e94e8e4..12d4817 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-use.html +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-use.html @@ -205,7 +205,7 @@ UnassignProcedure -Procedure that describe the unassignment of a single region. +Procedure that describes the unassignment of a single region. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/346adc37/devapidocs/org/apache/hadoop/hbase/master/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html index 9b1d336..86c040e 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html @@ -312,11 +312,11 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective -org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus +org.apache.hadoop.hbase.master.RegionState.State org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode +org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus +org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective org.apache.hadoop.hbase.master.MetricsMasterSourceFactoryImpl.FactoryStorage -org.apache.hadoop.hbase.master.RegionState.State http://git-wip-us.apache.org/repos/asf/hbase-site/blob/346adc37/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashException.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashException.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashException.html index 066581d..89087f7 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashException.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashException.html @@ -134,10 +134,11 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public class ServerCrashException +public class ServerCrashException extends HBaseIOException Passed as Exception by ServerCrashProcedure - notifying on-going RIT that server has failed. + notifying on-going RIT that server has failed. This exception is less an error-condition than + it is a signal to waiting procedures that they can now proceed. See Also: Serialized Form @@ -240,7 +241,7 @@ extends procId -private finallong procId +private finallong procId @@ -249,7 +250,7 @@ extends serverName -private finalServerName serverName +private finalServerName serverName @@ -266,7 +267,7 @@ extends ServerCrashException -publicServerCrashException(longprocId, +publicServerCrashException(longprocId, ServerNameserverName) Parameters: @@ -288,7 +289,7 @@ extends getMessage -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetMessage() +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetMessage() Overrides: http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getMessage--; title="class or interface in java.lang">getMessagein
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2aec596e/devapidocs/org/apache/hadoop/hbase/class-use/CellComparator.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/CellComparator.html b/devapidocs/org/apache/hadoop/hbase/class-use/CellComparator.html index c5d849a..6119d39 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/CellComparator.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/CellComparator.html @@ -253,12 +253,12 @@ int -RowIndexSeekerV1.compareKey(CellComparatorcomparator, +BufferedDataBlockEncoder.BufferedEncodedSeeker.compareKey(CellComparatorcomparator, Cellkey) int -BufferedDataBlockEncoder.BufferedEncodedSeeker.compareKey(CellComparatorcomparator, +RowIndexSeekerV1.compareKey(CellComparatorcomparator, Cellkey) @@ -270,27 +270,27 @@ DataBlockEncoder.EncodedSeeker -RowIndexCodecV1.createSeeker(CellComparatorcomparator, +CopyKeyDataBlockEncoder.createSeeker(CellComparatorcomparator, HFileBlockDecodingContextdecodingCtx) DataBlockEncoder.EncodedSeeker -CopyKeyDataBlockEncoder.createSeeker(CellComparatorcomparator, +PrefixKeyDeltaEncoder.createSeeker(CellComparatorcomparator, HFileBlockDecodingContextdecodingCtx) DataBlockEncoder.EncodedSeeker -DiffKeyDeltaEncoder.createSeeker(CellComparatorcomparator, +FastDiffDeltaEncoder.createSeeker(CellComparatorcomparator, HFileBlockDecodingContextdecodingCtx) DataBlockEncoder.EncodedSeeker -FastDiffDeltaEncoder.createSeeker(CellComparatorcomparator, +DiffKeyDeltaEncoder.createSeeker(CellComparatorcomparator, HFileBlockDecodingContextdecodingCtx) DataBlockEncoder.EncodedSeeker -PrefixKeyDeltaEncoder.createSeeker(CellComparatorcomparator, +RowIndexCodecV1.createSeeker(CellComparatorcomparator, HFileBlockDecodingContextdecodingCtx) @@ -328,9 +328,9 @@ -protected CellComparator -HFileWriterImpl.comparator -Key comparator. +private CellComparator +HFileBlockIndex.CellBasedKeyBlockIndexReader.comparator +Needed doing lookup on blocks. @@ -344,9 +344,9 @@ -private CellComparator -HFileBlockIndex.CellBasedKeyBlockIndexReader.comparator -Needed doing lookup on blocks. +protected CellComparator +HFileWriterImpl.comparator +Key comparator. @@ -530,8 +530,8 @@ StripeStoreFileManager.cellComparator -protected CellComparator -StripeMultiFileWriter.comparator +private CellComparator +StoreFileWriter.Builder.comparator protected CellComparator @@ -539,36 +539,36 @@ private CellComparator -Segment.comparator +AbstractMemStore.comparator private CellComparator -ScanInfo.comparator +Segment.comparator -private CellComparator -StoreFileWriter.Builder.comparator +protected CellComparator +HRegion.RegionScannerImpl.comparator protected CellComparator -HRegion.RegionScannerImpl.comparator +StripeMultiFileWriter.comparator private CellComparator -AbstractMemStore.comparator +ScanInfo.comparator protected CellComparator HStore.comparator -protected CellComparator -KeyValueHeap.KVScannerComparator.kvComparator - - private CellComparator DefaultStoreFileManager.kvComparator + +protected CellComparator +KeyValueHeap.KVScannerComparator.kvComparator + @@ -589,38 +589,38 @@ HRegion.getCellComparator() -CellComparator -StoreFileReader.getComparator() +(package private) CellComparator +StoreFileScanner.getComparator() protected CellComparator -Segment.getComparator() -Returns the Cell comparator used by this segment - +AbstractMemStore.getComparator() CellComparator -ScanInfo.getComparator() +StoreFileReader.getComparator() CellComparator KeyValueHeap.KVScannerComparator.getComparator() -CellComparator -Store.getComparator() +protected CellComparator +Segment.getComparator() +Returns the Cell comparator used by this segment + -protected CellComparator -AbstractMemStore.getComparator() +CellComparator +ScanInfo.getComparator() CellComparator -HStore.getComparator() +Store.getComparator() -(package private) CellComparator -StoreFileScanner.getComparator() +CellComparator +HStore.getComparator() @@ -663,21 +663,21 @@ protected void -DateTieredStoreEngine.createComponents(org.apache.hadoop.conf.Configurationconf, +StripeStoreEngine.createComponents(org.apache.hadoop.conf.Configurationconf, Storestore, -CellComparatorkvComparator) +CellComparatorcomparator) protected void -DefaultStoreEngine.createComponents(org.apache.hadoop.conf.Configurationconf, +DateTieredStoreEngine.createComponents(org.apache.hadoop.conf.Configurationconf, Storestore, CellComparatorkvComparator) protected void -StripeStoreEngine.createComponents(org.apache.hadoop.conf.Configurationconf, +DefaultStoreEngine.createComponents(org.apache.hadoop.conf.Configurationconf,
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1837997e/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MergeTableRegionProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MergeTableRegionProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MergeTableRegionProcedureBiConsumer.html index 35d5549..7f42873 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MergeTableRegionProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MergeTableRegionProcedureBiConsumer.html @@ -115,2816 +115,2814 @@ 107import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse; 108import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest; 109import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse; -110import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionRequest; -111import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionResponse; -112import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest; -113import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerResponse; -114import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest; -115import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationResponse; -116import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription; -117import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema; -118import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType; -119import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest; -120import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse; -121import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest; -122import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse; -123import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest; -124import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse; -125import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest; -126import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse; -127import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest; -128import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse; -129import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest; -130import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse; -131import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest; -132import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse; -133import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest; -134import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse; -135import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest; -136import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse; -137import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest; -138import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse; -139import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest; -140import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse; -141import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest; -142import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse; -143import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest; -144import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse; -145import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest; -146import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse; -147import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest; -148import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse; -149import
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a2b2dd19/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html index 5c95397..860416b 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html @@ -293,7944 +293,7962 @@ 285 final AtomicLong compactionsFailed = new AtomicLong(0L); 286 final AtomicLong compactionNumFilesCompacted = new AtomicLong(0L); 287 final AtomicLong compactionNumBytesCompacted = new AtomicLong(0L); -288 -289 private final WAL wal; -290 private final HRegionFileSystem fs; -291 protected final Configuration conf; -292 private final Configuration baseConf; -293 private final int rowLockWaitDuration; -294 static final int DEFAULT_ROWLOCK_WAIT_DURATION = 3; -295 -296 // The internal wait duration to acquire a lock before read/update -297 // from the region. It is not per row. The purpose of this wait time -298 // is to avoid waiting a long time while the region is busy, so that -299 // we can release the IPC handler soon enough to improve the -300 // availability of the region server. It can be adjusted by -301 // tuning configuration "hbase.busy.wait.duration". -302 final long busyWaitDuration; -303 static final long DEFAULT_BUSY_WAIT_DURATION = HConstants.DEFAULT_HBASE_RPC_TIMEOUT; -304 -305 // If updating multiple rows in one call, wait longer, -306 // i.e. waiting for busyWaitDuration * # of rows. However, -307 // we can limit the max multiplier. -308 final int maxBusyWaitMultiplier; -309 -310 // Max busy wait duration. There is no point to wait longer than the RPC -311 // purge timeout, when a RPC call will be terminated by the RPC engine. -312 final long maxBusyWaitDuration; -313 -314 // Max cell size. If nonzero, the maximum allowed size for any given cell -315 // in bytes -316 final long maxCellSize; -317 -318 // negative number indicates infinite timeout -319 static final long DEFAULT_ROW_PROCESSOR_TIMEOUT = 60 * 1000L; -320 final ExecutorService rowProcessorExecutor = Executors.newCachedThreadPool(); -321 -322 private final ConcurrentHashMapRegionScanner, Long scannerReadPoints; +288 final AtomicLong compactionsQueued = new AtomicLong(0L); +289 final AtomicLong flushesQueued = new AtomicLong(0L); +290 +291 private final WAL wal; +292 private final HRegionFileSystem fs; +293 protected final Configuration conf; +294 private final Configuration baseConf; +295 private final int rowLockWaitDuration; +296 static final int DEFAULT_ROWLOCK_WAIT_DURATION = 3; +297 +298 // The internal wait duration to acquire a lock before read/update +299 // from the region. It is not per row. The purpose of this wait time +300 // is to avoid waiting a long time while the region is busy, so that +301 // we can release the IPC handler soon enough to improve the +302 // availability of the region server. It can be adjusted by +303 // tuning configuration "hbase.busy.wait.duration". +304 final long busyWaitDuration; +305 static final long DEFAULT_BUSY_WAIT_DURATION = HConstants.DEFAULT_HBASE_RPC_TIMEOUT; +306 +307 // If updating multiple rows in one call, wait longer, +308 // i.e. waiting for busyWaitDuration * # of rows. However, +309 // we can limit the max multiplier. +310 final int maxBusyWaitMultiplier; +311 +312 // Max busy wait duration. There is no point to wait longer than the RPC +313 // purge timeout, when a RPC call will be terminated by the RPC engine. +314 final long maxBusyWaitDuration; +315 +316 // Max cell size. If nonzero, the maximum allowed size for any given cell +317 // in bytes +318 final long maxCellSize; +319 +320 // negative number indicates infinite timeout +321 static final long DEFAULT_ROW_PROCESSOR_TIMEOUT = 60 * 1000L; +322 final ExecutorService rowProcessorExecutor = Executors.newCachedThreadPool(); 323 -324 /** -325 * The sequence ID that was enLongAddered when this region was opened. -326 */ -327 private long openSeqNum = HConstants.NO_SEQNUM; -328 -329 /** -330 * The default setting for whether to enable on-demand CF loading for -331 * scan requests to this region. Requests can override it. -332 */ -333 private boolean isLoadingCfsOnDemandDefault = false; -334 -335 private final AtomicInteger majorInProgress = new AtomicInteger(0); -336 private final AtomicInteger minorInProgress = new AtomicInteger(0); -337 -338 // -339 // Context: During replay we want to ensure that we do not lose any data. So, we -340 // have to be conservative in how we replay wals. For each store, we calculate -341 // the maxSeqId up to which
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/21766f4a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/class-use/MinorCompactionScanQueryMatcher.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/class-use/MinorCompactionScanQueryMatcher.html b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/class-use/MinorCompactionScanQueryMatcher.html index 4d1e498..d53da57 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/class-use/MinorCompactionScanQueryMatcher.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/class-use/MinorCompactionScanQueryMatcher.html @@ -72,7 +72,49 @@ Uses of Classorg.apache.hadoop.hbase.regionserver.querymatcher.MinorCompactionScanQueryMatcher -No usage of org.apache.hadoop.hbase.regionserver.querymatcher.MinorCompactionScanQueryMatcher + + + + +Packages that use MinorCompactionScanQueryMatcher + +Package +Description + + + +org.apache.hadoop.hbase.regionserver.querymatcher + + + + + + + + + + +Uses of MinorCompactionScanQueryMatcher in org.apache.hadoop.hbase.regionserver.querymatcher + +Subclasses of MinorCompactionScanQueryMatcher in org.apache.hadoop.hbase.regionserver.querymatcher + +Modifier and Type +Class and Description + + + +class +IncludeAllCompactionQueryMatcher +A compaction query matcher that always return INCLUDE and drops nothing. + + + + + + + + + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/21766f4a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/class-use/NewVersionBehaviorTracker.DeleteVersionsNode.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/class-use/NewVersionBehaviorTracker.DeleteVersionsNode.html b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/class-use/NewVersionBehaviorTracker.DeleteVersionsNode.html new file mode 100644 index 000..ae2582c --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/class-use/NewVersionBehaviorTracker.DeleteVersionsNode.html @@ -0,0 +1,204 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +Uses of Class org.apache.hadoop.hbase.regionserver.querymatcher.NewVersionBehaviorTracker.DeleteVersionsNode (Apache HBase 3.0.0-SNAPSHOT API) + + + + + + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +Prev +Next + + +Frames +NoFrames + + +AllClasses + + + + + + + + + + +Uses of Classorg.apache.hadoop.hbase.regionserver.querymatcher.NewVersionBehaviorTracker.DeleteVersionsNode + + + + + +Packages that use NewVersionBehaviorTracker.DeleteVersionsNode + +Package +Description + + + +org.apache.hadoop.hbase.regionserver.querymatcher + + + +org.apache.hadoop.hbase.security.visibility + + + + + + + + + + +Uses of NewVersionBehaviorTracker.DeleteVersionsNode in org.apache.hadoop.hbase.regionserver.querymatcher + +Fields in org.apache.hadoop.hbase.regionserver.querymatcher with type parameters of type NewVersionBehaviorTracker.DeleteVersionsNode + +Modifier and Type +Field and Description + + + +protected http://docs.oracle.com/javase/8/docs/api/java/util/NavigableMap.html?is-external=true; title="class or interface in java.util">NavigableMaphttp://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">Long,NewVersionBehaviorTracker.DeleteVersionsNode +NewVersionBehaviorTracker.delColMap + + +protected http://docs.oracle.com/javase/8/docs/api/java/util/NavigableMap.html?is-external=true; title="class or interface in java.util">NavigableMaphttp://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">Long,NewVersionBehaviorTracker.DeleteVersionsNode +NewVersionBehaviorTracker.delFamMap + + + + +Methods in org.apache.hadoop.hbase.regionserver.querymatcher that return NewVersionBehaviorTracker.DeleteVersionsNode + +Modifier and Type +Method and Description + + + +protected NewVersionBehaviorTracker.DeleteVersionsNode +NewVersionBehaviorTracker.DeleteVersionsNode.getDeepCopy() + + + + + + + +Uses of NewVersionBehaviorTracker.DeleteVersionsNode in org.apache.hadoop.hbase.security.visibility + +Subclasses of
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2d5075d7/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html index 22817f5..582d47a 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html @@ -692,19 +692,19 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.regionserver.ScanType -org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection -org.apache.hadoop.hbase.regionserver.BloomType +org.apache.hadoop.hbase.regionserver.RegionOpeningState +org.apache.hadoop.hbase.regionserver.ScannerContext.NextState org.apache.hadoop.hbase.regionserver.Region.FlushResult.Result +org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status org.apache.hadoop.hbase.regionserver.Region.Operation +org.apache.hadoop.hbase.regionserver.BloomType +org.apache.hadoop.hbase.regionserver.FlushType +org.apache.hadoop.hbase.regionserver.ScanType org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope -org.apache.hadoop.hbase.regionserver.ScannerContext.NextState -org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType org.apache.hadoop.hbase.regionserver.MemStoreCompactor.Action -org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status -org.apache.hadoop.hbase.regionserver.RegionOpeningState -org.apache.hadoop.hbase.regionserver.FlushType +org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection +org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage org.apache.hadoop.hbase.regionserver.StoreScanner.StoreScannerCompactionRace http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2d5075d7/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html b/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html index 7cba16c..33cb6b6 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html @@ -248,8 +248,8 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.regionserver.wal.RingBufferTruck.Type org.apache.hadoop.hbase.regionserver.wal.ProtobufLogReader.WALHdrResult +org.apache.hadoop.hbase.regionserver.wal.RingBufferTruck.Type http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2d5075d7/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html index 1185ae8..2680226 100644 --- a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html @@ -214,8 +214,8 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.replication.regionserver.ReplicationSourceShipper.WorkerState org.apache.hadoop.hbase.replication.regionserver.MetricsReplicationSourceFactoryImpl.SourceHolder +org.apache.hadoop.hbase.replication.regionserver.ReplicationSourceShipper.WorkerState
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0383a9c2/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.html index 6c200a1..e6f8c2e 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.html @@ -1350,726 +1350,719 @@ 1342 } 1343 1344 @Override -1345 public MasterProtos.DispatchMergingRegionsResponse dispatchMergingRegions( -1346 RpcController controller, MasterProtos.DispatchMergingRegionsRequest request) -1347 throws ServiceException { -1348return stub.dispatchMergingRegions(controller, request); -1349 } -1350 -1351 @Override -1352 public MasterProtos.AssignRegionResponse assignRegion(RpcController controller, -1353 MasterProtos.AssignRegionRequest request) throws ServiceException { -1354return stub.assignRegion(controller, request); -1355 } -1356 -1357 @Override -1358 public MasterProtos.UnassignRegionResponse unassignRegion(RpcController controller, -1359 MasterProtos.UnassignRegionRequest request) throws ServiceException { -1360return stub.unassignRegion(controller, request); -1361 } -1362 -1363 @Override -1364 public MasterProtos.OfflineRegionResponse offlineRegion(RpcController controller, -1365 MasterProtos.OfflineRegionRequest request) throws ServiceException { -1366return stub.offlineRegion(controller, request); -1367 } -1368 -1369 @Override -1370 public MasterProtos.SplitTableRegionResponse splitRegion(RpcController controller, -1371 MasterProtos.SplitTableRegionRequest request) throws ServiceException { -1372return stub.splitRegion(controller, request); -1373 } -1374 -1375 @Override -1376 public MasterProtos.DeleteTableResponse deleteTable(RpcController controller, -1377 MasterProtos.DeleteTableRequest request) throws ServiceException { -1378return stub.deleteTable(controller, request); -1379 } -1380 -1381 @Override -1382 public MasterProtos.TruncateTableResponse truncateTable(RpcController controller, -1383 MasterProtos.TruncateTableRequest request) throws ServiceException { -1384return stub.truncateTable(controller, request); -1385 } -1386 -1387 @Override -1388 public MasterProtos.EnableTableResponse enableTable(RpcController controller, -1389 MasterProtos.EnableTableRequest request) throws ServiceException { -1390return stub.enableTable(controller, request); -1391 } -1392 -1393 @Override -1394 public MasterProtos.DisableTableResponse disableTable(RpcController controller, -1395 MasterProtos.DisableTableRequest request) throws ServiceException { -1396return stub.disableTable(controller, request); -1397 } -1398 -1399 @Override -1400 public MasterProtos.ModifyTableResponse modifyTable(RpcController controller, -1401 MasterProtos.ModifyTableRequest request) throws ServiceException { -1402return stub.modifyTable(controller, request); -1403 } -1404 -1405 @Override -1406 public MasterProtos.CreateTableResponse createTable(RpcController controller, -1407 MasterProtos.CreateTableRequest request) throws ServiceException { -1408return stub.createTable(controller, request); -1409 } -1410 -1411 @Override -1412 public MasterProtos.ShutdownResponse shutdown(RpcController controller, -1413 MasterProtos.ShutdownRequest request) throws ServiceException { -1414return stub.shutdown(controller, request); -1415 } -1416 -1417 @Override -1418 public MasterProtos.StopMasterResponse stopMaster(RpcController controller, -1419 MasterProtos.StopMasterRequest request) throws ServiceException { -1420return stub.stopMaster(controller, request); +1345 public MasterProtos.AssignRegionResponse assignRegion(RpcController controller, +1346 MasterProtos.AssignRegionRequest request) throws ServiceException { +1347return stub.assignRegion(controller, request); +1348 } +1349 +1350 @Override +1351 public MasterProtos.UnassignRegionResponse unassignRegion(RpcController controller, +1352 MasterProtos.UnassignRegionRequest request) throws ServiceException { +1353return stub.unassignRegion(controller, request); +1354 } +1355 +1356 @Override +1357 public MasterProtos.OfflineRegionResponse offlineRegion(RpcController controller, +1358 MasterProtos.OfflineRegionRequest request) throws
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f391bcef/devapidocs/index-all.html -- diff --git a/devapidocs/index-all.html b/devapidocs/index-all.html index 8b8f33b..ad2b82a 100644 --- a/devapidocs/index-all.html +++ b/devapidocs/index-all.html @@ -778,6 +778,8 @@ Action(Row, int) - Constructor for class org.apache.hadoop.hbase.client.Action +Action(Row, int, int) - Constructor for class org.apache.hadoop.hbase.client.Action + Action(Action, int) - Constructor for class org.apache.hadoop.hbase.client.Action Creates an action for a particular replica from original action. @@ -8929,7 +8931,7 @@ This class is used to unify HTable calls with AsyncProcess Framework. -CancellableRegionServerCallable(Connection, TableName, byte[], RpcController, int, RetryingTimeTracker) - Constructor for class org.apache.hadoop.hbase.client.CancellableRegionServerCallable +CancellableRegionServerCallable(Connection, TableName, byte[], RpcController, int, RetryingTimeTracker, int) - Constructor for class org.apache.hadoop.hbase.client.CancellableRegionServerCallable cancellationCbs - Variable in class org.apache.hadoop.hbase.ipc.HBaseRpcControllerImpl @@ -12006,7 +12008,7 @@ A RegionServerCallable set to use the Client protocol. -ClientServiceCallable(Connection, TableName, byte[], RpcController) - Constructor for class org.apache.hadoop.hbase.client.ClientServiceCallable +ClientServiceCallable(Connection, TableName, byte[], RpcController, int) - Constructor for class org.apache.hadoop.hbase.client.ClientServiceCallable ClientSideRegionScanner - Class in org.apache.hadoop.hbase.client @@ -40151,6 +40153,8 @@ getMaxPoolSize() - Method in interface org.apache.hadoop.hbase.thrift.IncrementCoalescerMBean +getMaxPriority() - Method in class org.apache.hadoop.hbase.client.RowMutations + getMaxProcId() - Method in class org.apache.hadoop.hbase.procedure2.store.wal.ProcedureWALFile getMaxProcId() - Method in class org.apache.hadoop.hbase.procedure2.store.wal.ProcedureWALFormatReader.WalProcedureMap @@ -43284,6 +43288,14 @@ Get the user name from a principal +getPriority() - Method in class org.apache.hadoop.hbase.client.Action + +getPriority() - Method in class org.apache.hadoop.hbase.client.MultiAction + +getPriority() - Method in class org.apache.hadoop.hbase.client.OperationWithAttributes + +getPriority() - Method in class org.apache.hadoop.hbase.client.RegionServerCallable + getPriority() - Method in interface org.apache.hadoop.hbase.client.TableDescriptor getPriority() - Method in class org.apache.hadoop.hbase.client.TableDescriptorBuilder.ModifyableTableDescriptor @@ -53729,20 +53741,21 @@ HColumnDescriptor(String) - Constructor for class org.apache.hadoop.hbase.HColumnDescriptor -Deprecated. -Construct a column descriptor specifying only the family name - The other attributes are defaulted. +Deprecated. +use ColumnFamilyDescriptorBuilder.of(String) + HColumnDescriptor(byte[]) - Constructor for class org.apache.hadoop.hbase.HColumnDescriptor -Deprecated. -Construct a column descriptor specifying only the family name - The other attributes are defaulted. +Deprecated. +use ColumnFamilyDescriptorBuilder.of(byte[]) + HColumnDescriptor(HColumnDescriptor) - Constructor for class org.apache.hadoop.hbase.HColumnDescriptor -Deprecated. -Constructor. +Deprecated. +use ColumnFamilyDescriptorBuilder.copy(ColumnFamilyDescriptor) + HColumnDescriptor(HColumnDescriptor, boolean) - Constructor for class org.apache.hadoop.hbase.HColumnDescriptor @@ -72247,7 +72260,7 @@ Callable that handles the multi method call going against a single regionserver; i.e. -MultiServerCallable(ClusterConnection, TableName, ServerName, MultiAction, RpcController, int, RetryingTimeTracker) - Constructor for class org.apache.hadoop.hbase.client.MultiServerCallable +MultiServerCallable(ClusterConnection, TableName, ServerName, MultiAction, RpcController, int, RetryingTimeTracker, int) - Constructor for class org.apache.hadoop.hbase.client.MultiServerCallable multiServersPerHost - Variable in class org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster @@ -74229,7 +74242,7 @@ Implementations make an rpc call against a RegionService via a protobuf Service. -NoncedRegionServerCallable(Connection, TableName, byte[], HBaseRpcController) - Constructor for class org.apache.hadoop.hbase.client.NoncedRegionServerCallable +NoncedRegionServerCallable(Connection, TableName, byte[], HBaseRpcController, int) - Constructor for class org.apache.hadoop.hbase.client.NoncedRegionServerCallable nonceGenerator - Variable in class org.apache.hadoop.hbase.client.AsyncConnectionImpl @@ -74400,10 +74413,7 @@ noReadahead - Variable in class org.apache.hadoop.hbase.regionserver.HStoreFile NORMAL_QOS - Static variable in class org.apache.hadoop.hbase.HConstants - -QOS attributes:
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ca5b0275/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html index 2ca99ba..2227828 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html @@ -123,7 +123,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.LimitedPrivate(value="Tools") -public class HRegionServer +public class HRegionServer extends HasThread implements RegionServerServices, LastSequenceId, ConfigurationObserver HRegionServer makes a set of HRegions available to clients. It checks in with @@ -1534,7 +1534,7 @@ implements INIT_PAUSE_TIME_MS -private static finalint INIT_PAUSE_TIME_MS +private static finalint INIT_PAUSE_TIME_MS See Also: Constant Field Values @@ -1547,7 +1547,7 @@ implements REGION_LOCK_AWAIT_TIME_SEC -public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String REGION_LOCK_AWAIT_TIME_SEC +public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String REGION_LOCK_AWAIT_TIME_SEC See Also: Constant Field Values @@ -1560,7 +1560,7 @@ implements DEFAULT_REGION_LOCK_AWAIT_TIME_SEC -public static finalint DEFAULT_REGION_LOCK_AWAIT_TIME_SEC +public static finalint DEFAULT_REGION_LOCK_AWAIT_TIME_SEC See Also: Constant Field Values @@ -1573,7 +1573,7 @@ implements LOG -private static finalorg.apache.commons.logging.Log LOG +private static finalorg.apache.commons.logging.Log LOG @@ -1582,7 +1582,7 @@ implements TEST_SKIP_REPORTING_TRANSITION -public staticboolean TEST_SKIP_REPORTING_TRANSITION +public staticboolean TEST_SKIP_REPORTING_TRANSITION For testing only! Set to true to skip notifying region assignment to master . @@ -1592,7 +1592,7 @@ implements OPEN -protected static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String OPEN +protected static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String OPEN See Also: Constant Field Values @@ -1605,7 +1605,7 @@ implements CLOSE -protected static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String CLOSE +protected static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String CLOSE See Also: Constant Field Values @@ -1618,7 +1618,7 @@ implements regionsInTransitionInRS -protected finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true; title="class or interface in java.util.concurrent">ConcurrentMapbyte[],http://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean regionsInTransitionInRS +protected finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true; title="class or interface in java.util.concurrent">ConcurrentMapbyte[],http://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean regionsInTransitionInRS @@ -1627,7 +1627,7 @@ implements cacheFlusher -protectedMemStoreFlusher cacheFlusher +protectedMemStoreFlusher cacheFlusher @@ -1636,7 +1636,7 @@ implements hMemManager -protectedHeapMemoryManager hMemManager +protectedHeapMemoryManager hMemManager @@ -1645,7 +1645,7 @@ implements initLatch -protectedhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CountDownLatch.html?is-external=true; title="class or interface in java.util.concurrent">CountDownLatch initLatch +protectedhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CountDownLatch.html?is-external=true; title="class or interface in java.util.concurrent">CountDownLatch initLatch @@ -1654,7 +1654,7 @@ implements clusterConnection -protectedClusterConnection clusterConnection +protectedClusterConnection clusterConnection Cluster connection to be shared by services. Initialized at server startup and closed when server shuts down. Clients must never close it explicitly. @@ -1666,7 +1666,7 @@ implements metaTableLocator -protectedMetaTableLocator metaTableLocator +protectedMetaTableLocator metaTableLocator @@ -1675,7 +1675,7 @@ implements recoveringRegionWatcher -privateRecoveringRegionWatcher recoveringRegionWatcher +privateRecoveringRegionWatcher
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/9eba7fcf/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html index 504e470..38667c0 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html @@ -2866,5375 +2866,5371 @@ 2858checkResources(); 2859 startRegionOperation(Operation.DELETE); 2860try { -2861 delete.getRow(); -2862 // All edits for the given row (across all column families) must happen atomically. -2863 doBatchMutate(delete); -2864} finally { -2865 closeRegionOperation(Operation.DELETE); -2866} -2867 } -2868 -2869 /** -2870 * Row needed by below method. -2871 */ -2872 private static final byte [] FOR_UNIT_TESTS_ONLY = Bytes.toBytes("ForUnitTestsOnly"); -2873 -2874 /** -2875 * This is used only by unit tests. Not required to be a public API. -2876 * @param familyMap map of family to edits for the given family. -2877 * @throws IOException -2878 */ -2879 void delete(NavigableMapbyte[], ListCell familyMap, -2880 Durability durability) throws IOException { -2881Delete delete = new Delete(FOR_UNIT_TESTS_ONLY); -2882 delete.setFamilyCellMap(familyMap); -2883delete.setDurability(durability); -2884doBatchMutate(delete); -2885 } -2886 -2887 @Override -2888 public void prepareDeleteTimestamps(Mutation mutation, Mapbyte[], ListCell familyMap, -2889 byte[] byteNow) throws IOException { -2890for (Map.Entrybyte[], ListCell e : familyMap.entrySet()) { -2891 -2892 byte[] family = e.getKey(); -2893 ListCell cells = e.getValue(); -2894 assert cells instanceof RandomAccess; -2895 -2896 Mapbyte[], Integer kvCount = new TreeMap(Bytes.BYTES_COMPARATOR); -2897 int listSize = cells.size(); -2898 for (int i=0; i listSize; i++) { -2899Cell cell = cells.get(i); -2900// Check if time is LATEST, change to time of most recent addition if so -2901// This is expensive. -2902if (cell.getTimestamp() == HConstants.LATEST_TIMESTAMP CellUtil.isDeleteType(cell)) { -2903 byte[] qual = CellUtil.cloneQualifier(cell); -2904 if (qual == null) qual = HConstants.EMPTY_BYTE_ARRAY; -2905 -2906 Integer count = kvCount.get(qual); -2907 if (count == null) { -2908kvCount.put(qual, 1); -2909 } else { -2910kvCount.put(qual, count + 1); -2911 } -2912 count = kvCount.get(qual); -2913 -2914 Get get = new Get(CellUtil.cloneRow(cell)); -2915 get.setMaxVersions(count); -2916 get.addColumn(family, qual); -2917 if (coprocessorHost != null) { -2918if (!coprocessorHost.prePrepareTimeStampForDeleteVersion(mutation, cell, -2919byteNow, get)) { -2920 updateDeleteLatestVersionTimeStamp(cell, get, count, byteNow); -2921} -2922 } else { -2923 updateDeleteLatestVersionTimeStamp(cell, get, count, byteNow); -2924 } -2925} else { -2926 CellUtil.updateLatestStamp(cell, byteNow, 0); -2927} -2928 } -2929} -2930 } -2931 -2932 void updateDeleteLatestVersionTimeStamp(Cell cell, Get get, int count, byte[] byteNow) -2933 throws IOException { -2934ListCell result = get(get, false); -2935 -2936if (result.size() count) { -2937 // Nothing to delete -2938 CellUtil.updateLatestStamp(cell, byteNow, 0); -2939 return; -2940} -2941if (result.size() count) { -2942 throw new RuntimeException("Unexpected size: " + result.size()); -2943} -2944Cell getCell = result.get(count - 1); -2945CellUtil.setTimestamp(cell, getCell.getTimestamp()); -2946 } -2947 -2948 @Override -2949 public void put(Put put) throws IOException { -2950checkReadOnly(); -2951 -2952// Do a rough check that we have resources to accept a write. The check is -2953// 'rough' in that between the resource check and the call to obtain a -2954// read lock, resources may run out. For now, the thought is that this -2955// will be extremely rare; we'll deal with it when it happens. -2956checkResources(); -2957 startRegionOperation(Operation.PUT); -2958try { -2959 // All edits for the given row (across all column families) must happen atomically. -2960 doBatchMutate(put); -2961} finally { -2962 closeRegionOperation(Operation.PUT); -2963} -2964 } -2965 -2966 /** -2967 * Struct-like class that tracks the progress of a batch operation, -2968 * accumulating status
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/17128d27/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html index 2fdac6a..def5ec9 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html @@ -1192,686 +1192,686 @@ 1184 * 1185 * @param regionName region to split 1186 * @throws IOException if a remote or network exception occurs -1187 */ -1188 void splitRegion(final byte[] regionName) throws IOException; -1189 -1190 /** -1191 * Split a table. Asynchronous operation. -1192 * -1193 * @param tableName table to split -1194 * @param splitPoint the explicit position to split on -1195 * @throws IOException if a remote or network exception occurs -1196 */ -1197 void split(final TableName tableName, final byte[] splitPoint) -1198throws IOException; -1199 -1200 /** -1201 * Split an individual region. Asynchronous operation. -1202 * -1203 * @param regionName region to split -1204 * @param splitPoint the explicit position to split on -1205 * @throws IOException if a remote or network exception occurs -1206 */ -1207 void splitRegion(final byte[] regionName, final byte[] splitPoint) -1208throws IOException; -1209 -1210 /** -1211 * Modify an existing table, more IRB friendly version. -1212 * -1213 * @param tableName name of table. -1214 * @param htd modified description of the table -1215 * @throws IOException if a remote or network exception occurs -1216 * @deprecated since 2.0 version and will be removed in 3.0 version. -1217 * use {@link #modifyTable(TableDescriptor)} -1218 */ -1219 @Deprecated -1220 void modifyTable(final TableName tableName, final HTableDescriptor htd) -1221 throws IOException; +1187 * @deprecated Since 2.0. Will be removed in 3.0. Use +1188 * {@link #splitRegionAsync(byte[], byte[])} instead. +1189 */ +1190 void splitRegion(final byte[] regionName) throws IOException; +1191 +1192 /** +1193 * Split a table. Asynchronous operation. +1194 * +1195 * @param tableName table to split +1196 * @param splitPoint the explicit position to split on +1197 * @throws IOException if a remote or network exception occurs +1198 */ +1199 void split(final TableName tableName, final byte[] splitPoint) +1200throws IOException; +1201 +1202 /** +1203 * Split an individual region. Asynchronous operation. +1204 * +1205 * @param regionName region to split +1206 * @param splitPoint the explicit position to split on +1207 * @throws IOException if a remote or network exception occurs +1208 * @deprecated Since 2.0. Will be removed in 3.0. Use +1209 * {@link #splitRegionAsync(byte[], byte[])} instead. +1210 */ +1211 void splitRegion(final byte[] regionName, final byte[] splitPoint) +1212throws IOException; +1213 +1214 /** +1215 * Split an individual region. Asynchronous operation. +1216 * @param regionName region to split +1217 * @param splitPoint the explicit position to split on +1218 * @throws IOException if a remote or network exception occurs +1219 */ +1220 FutureVoid splitRegionAsync(byte[] regionName, byte[] splitPoint) +1221throws IOException; 1222 1223 /** 1224 * Modify an existing table, more IRB friendly version. 1225 * -1226 * @param td modified description of the table -1227 * @throws IOException if a remote or network exception occurs -1228 */ -1229 void modifyTable(final TableDescriptor td) throws IOException; -1230 -1231 /** -1232 * Modify an existing table, more IRB friendly version. Asynchronous operation. This means that -1233 * it may be a while before your schema change is updated across all of the table. -1234 * You can use Future.get(long, TimeUnit) to wait on the operation to complete. -1235 * It may throw ExecutionException if there was an error while executing the operation -1236 * or TimeoutException in case the wait timeout was not long enough to allow the -1237 * operation to complete. +1226 * @param tableName name of table. +1227 * @param htd modified description of the table +1228 * @throws IOException if a remote or network exception occurs +1229 * @deprecated since 2.0 version and will be removed in 3.0 version. +1230 * use {@link #modifyTable(TableDescriptor)} +1231 */ +1232 @Deprecated +1233 void modifyTable(final TableName tableName, final HTableDescriptor htd) +1234 throws IOException; +1235 +1236 /** +1237 * Modify an existing table, more IRB friendly version. 1238 * -1239 * @param tableName name of table. -1240 * @param htd modified description of the table -1241 * @throws IOException if a remote or network
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2777c693/devapidocs/src-html/org/apache/hadoop/hbase/ipc/AbstractRpcClient.BlockingRpcChannelImplementation.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/AbstractRpcClient.BlockingRpcChannelImplementation.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/AbstractRpcClient.BlockingRpcChannelImplementation.html index 30d5dd8..beacb44 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/AbstractRpcClient.BlockingRpcChannelImplementation.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/AbstractRpcClient.BlockingRpcChannelImplementation.html @@ -109,497 +109,499 @@ 101 private static final ScheduledExecutorService IDLE_CONN_SWEEPER = Executors 102 .newScheduledThreadPool(1, Threads.newDaemonThreadFactory("Idle-Rpc-Conn-Sweeper")); 103 -104 protected final static MapKind, TokenSelector? extends TokenIdentifier TOKEN_HANDLERS = new HashMap(); -105 -106 static { -107 TOKEN_HANDLERS.put(Kind.HBASE_AUTH_TOKEN, new AuthenticationTokenSelector()); -108 } -109 -110 protected boolean running = true; // if client runs +104 @edu.umd.cs.findbugs.annotations.SuppressWarnings(value="MS_MUTABLE_COLLECTION_PKGPROTECT", +105 justification="the rest of the system which live in the different package can use") +106 protected final static MapKind, TokenSelector? extends TokenIdentifier TOKEN_HANDLERS = new HashMap(); +107 +108 static { +109 TOKEN_HANDLERS.put(Kind.HBASE_AUTH_TOKEN, new AuthenticationTokenSelector()); +110 } 111 -112 protected final Configuration conf; -113 protected final String clusterId; -114 protected final SocketAddress localAddr; -115 protected final MetricsConnection metrics; -116 -117 protected final UserProvider userProvider; -118 protected final CellBlockBuilder cellBlockBuilder; -119 -120 protected final int minIdleTimeBeforeClose; // if the connection is idle for more than this -121 // time (in ms), it will be closed at any moment. -122 protected final int maxRetries; // the max. no. of retries for socket connections -123 protected final long failureSleep; // Time to sleep before retry on failure. -124 protected final boolean tcpNoDelay; // if T then disable Nagle's Algorithm -125 protected final boolean tcpKeepAlive; // if T then use keepalives -126 protected final Codec codec; -127 protected final CompressionCodec compressor; -128 protected final boolean fallbackAllowed; -129 -130 protected final FailedServers failedServers; +112 protected boolean running = true; // if client runs +113 +114 protected final Configuration conf; +115 protected final String clusterId; +116 protected final SocketAddress localAddr; +117 protected final MetricsConnection metrics; +118 +119 protected final UserProvider userProvider; +120 protected final CellBlockBuilder cellBlockBuilder; +121 +122 protected final int minIdleTimeBeforeClose; // if the connection is idle for more than this +123 // time (in ms), it will be closed at any moment. +124 protected final int maxRetries; // the max. no. of retries for socket connections +125 protected final long failureSleep; // Time to sleep before retry on failure. +126 protected final boolean tcpNoDelay; // if T then disable Nagle's Algorithm +127 protected final boolean tcpKeepAlive; // if T then use keepalives +128 protected final Codec codec; +129 protected final CompressionCodec compressor; +130 protected final boolean fallbackAllowed; 131 -132 protected final int connectTO; -133 protected final int readTO; -134 protected final int writeTO; -135 -136 protected final PoolMapConnectionId, T connections; +132 protected final FailedServers failedServers; +133 +134 protected final int connectTO; +135 protected final int readTO; +136 protected final int writeTO; 137 -138 private final AtomicInteger callIdCnt = new AtomicInteger(0); +138 protected final PoolMapConnectionId, T connections; 139 -140 private final ScheduledFuture? cleanupIdleConnectionTask; +140 private final AtomicInteger callIdCnt = new AtomicInteger(0); 141 -142 private int maxConcurrentCallsPerServer; +142 private final ScheduledFuture? cleanupIdleConnectionTask; 143 -144 private static final LoadingCacheInetSocketAddress, AtomicInteger concurrentCounterCache = -145 CacheBuilder.newBuilder().expireAfterAccess(1, TimeUnit.HOURS). -146 build(new CacheLoaderInetSocketAddress, AtomicInteger() { -147@Override public AtomicInteger load(InetSocketAddress key) throws Exception { -148 return new AtomicInteger(0); -149} -150 }); -151 -152 /** -153 * Construct an IPC client for the cluster codeclusterId/code -154 * @param conf configuration -155 * @param clusterId the cluster id -156 * @param localAddr client socket bind address. -157 * @param metrics the
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/90c7dfe4/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html index 75db22d..99a09f9 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html @@ -37,2710 +37,2816 @@ 029import java.util.List; 030import java.util.Map; 031import java.util.Optional; -032import java.util.concurrent.CompletableFuture; -033import java.util.concurrent.TimeUnit; -034import java.util.concurrent.atomic.AtomicReference; -035import java.util.function.BiConsumer; -036import java.util.regex.Pattern; -037import java.util.stream.Collectors; -038 -039import com.google.common.annotations.VisibleForTesting; -040 -041import io.netty.util.Timeout; -042import io.netty.util.TimerTask; -043 -044import java.util.stream.Stream; -045 -046import org.apache.commons.io.IOUtils; -047import org.apache.commons.logging.Log; -048import org.apache.commons.logging.LogFactory; -049import org.apache.hadoop.hbase.ClusterStatus; -050import org.apache.hadoop.hbase.HRegionInfo; -051import org.apache.hadoop.hbase.HRegionLocation; -052import org.apache.hadoop.hbase.MetaTableAccessor; -053import org.apache.hadoop.hbase.MetaTableAccessor.QueryType; -054import org.apache.hadoop.hbase.NotServingRegionException; -055import org.apache.hadoop.hbase.ProcedureInfo; -056import org.apache.hadoop.hbase.RegionLoad; -057import org.apache.hadoop.hbase.RegionLocations; -058import org.apache.hadoop.hbase.ServerName; -059import org.apache.hadoop.hbase.NamespaceDescriptor; -060import org.apache.hadoop.hbase.HConstants; -061import org.apache.hadoop.hbase.TableExistsException; -062import org.apache.hadoop.hbase.TableName; -063import org.apache.hadoop.hbase.AsyncMetaTableAccessor; -064import org.apache.hadoop.hbase.TableNotDisabledException; -065import org.apache.hadoop.hbase.TableNotEnabledException; -066import org.apache.hadoop.hbase.TableNotFoundException; -067import org.apache.hadoop.hbase.UnknownRegionException; -068import org.apache.hadoop.hbase.classification.InterfaceAudience; -069import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder; -070import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder; -071import org.apache.hadoop.hbase.client.Scan.ReadType; -072import org.apache.hadoop.hbase.client.replication.ReplicationSerDeHelper; -073import org.apache.hadoop.hbase.client.replication.TableCFs; -074import org.apache.hadoop.hbase.exceptions.DeserializationException; -075import org.apache.hadoop.hbase.ipc.HBaseRpcController; -076import org.apache.hadoop.hbase.quotas.QuotaFilter; -077import org.apache.hadoop.hbase.quotas.QuotaSettings; -078import org.apache.hadoop.hbase.quotas.QuotaTableUtil; -079import org.apache.hadoop.hbase.replication.ReplicationException; -080import org.apache.hadoop.hbase.replication.ReplicationPeerConfig; -081import org.apache.hadoop.hbase.replication.ReplicationPeerDescription; -082import org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback; -083import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -084import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter; -085import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService; -086import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest; -087import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse; -088import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest; -089import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse; -090import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest; -091import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionResponse; -092import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionRequest; -093import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionResponse; -094import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest; -095import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse; -096import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequest; -097import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse; -098import
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0821e51a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteNamespaceProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteNamespaceProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteNamespaceProcedureBiConsumer.html index 71844ce..75db22d 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteNamespaceProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteNamespaceProcedureBiConsumer.html @@ -105,2564 +105,2642 @@ 097import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse; 098import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionRequest; 099import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionResponse; -100import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription; -101import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema; -102import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType; -103import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest; -104import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse; -105import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest; -106import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse; -107import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest; -108import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse; -109import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest; -110import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse; -111import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest; -112import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse; -113import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest; -114import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse; -115import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest; -116import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse; -117import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest; -118import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse; -119import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest; -120import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse; -121import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest; -122import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse; -123import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest; -124import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse; -125import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest; -126import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse; -127import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest; -128import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse; -129import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest; -130import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse; -131import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest; -132import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse; -133import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest; -134import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse; -135import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest; -136import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse; -137import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest; -138import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse; -139import
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/9fb0764b/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html b/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html index 2def063..4e1f2ed 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html @@ -62,854 +62,860 @@ 054import org.apache.hadoop.hbase.client.RowMutations; 055import org.apache.hadoop.hbase.client.Scan; 056import org.apache.hadoop.hbase.client.Table; -057import org.apache.hadoop.hbase.client.coprocessor.Batch; -058import org.apache.hadoop.hbase.client.coprocessor.Batch.Callback; -059import org.apache.hadoop.hbase.client.metrics.ScanMetrics; -060import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; -061import org.apache.hadoop.hbase.io.TimeRange; -062import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel; -063import org.apache.hadoop.hbase.rest.Constants; -064import org.apache.hadoop.hbase.rest.model.CellModel; -065import org.apache.hadoop.hbase.rest.model.CellSetModel; -066import org.apache.hadoop.hbase.rest.model.RowModel; -067import org.apache.hadoop.hbase.rest.model.ScannerModel; -068import org.apache.hadoop.hbase.rest.model.TableSchemaModel; -069import org.apache.hadoop.hbase.util.Bytes; -070import org.apache.hadoop.util.StringUtils; -071 -072import com.google.protobuf.Descriptors; -073import com.google.protobuf.Message; -074import com.google.protobuf.Service; -075import com.google.protobuf.ServiceException; -076 -077/** -078 * HTable interface to remote tables accessed via REST gateway -079 */ -080@InterfaceAudience.Public -081public class RemoteHTable implements Table { -082 -083 private static final Log LOG = LogFactory.getLog(RemoteHTable.class); -084 -085 final Client client; -086 final Configuration conf; -087 final byte[] name; -088 final int maxRetries; -089 final long sleepTime; -090 -091 @SuppressWarnings("rawtypes") -092 protected String buildRowSpec(final byte[] row, final Map familyMap, -093 final long startTime, final long endTime, final int maxVersions) { -094StringBuffer sb = new StringBuffer(); -095sb.append('/'); -096sb.append(Bytes.toString(name)); -097sb.append('/'); -098sb.append(toURLEncodedBytes(row)); -099Set families = familyMap.entrySet(); -100if (families != null) { -101 Iterator i = familyMap.entrySet().iterator(); -102 sb.append('/'); -103 while (i.hasNext()) { -104Map.Entry e = (Map.Entry)i.next(); -105Collection quals = (Collection)e.getValue(); -106if (quals == null || quals.isEmpty()) { -107 // this is an unqualified family. append the family name and NO ':' -108 sb.append(toURLEncodedBytes((byte[])e.getKey())); -109} else { -110 Iterator ii = quals.iterator(); -111 while (ii.hasNext()) { -112 sb.append(toURLEncodedBytes((byte[])e.getKey())); -113sb.append(':'); -114Object o = ii.next(); -115// Puts use byte[] but Deletes use KeyValue -116if (o instanceof byte[]) { -117 sb.append(toURLEncodedBytes((byte[])o)); -118} else if (o instanceof KeyValue) { -119 sb.append(toURLEncodedBytes(CellUtil.cloneQualifier((KeyValue)o))); -120} else { -121 throw new RuntimeException("object type not handled"); -122} -123if (ii.hasNext()) { -124 sb.append(','); -125} -126 } -127} -128if (i.hasNext()) { -129 sb.append(','); -130} -131 } -132} -133if (startTime = 0 endTime != Long.MAX_VALUE) { -134 sb.append('/'); -135 sb.append(startTime); -136 if (startTime != endTime) { -137sb.append(','); -138sb.append(endTime); -139 } -140} else if (endTime != Long.MAX_VALUE) { -141 sb.append('/'); -142 sb.append(endTime); -143} -144if (maxVersions 1) { -145 sb.append("?v="); -146 sb.append(maxVersions); -147} -148return sb.toString(); -149 } -150 -151 protected String buildMultiRowSpec(final byte[][] rows, int maxVersions) { -152StringBuilder sb = new StringBuilder(); -153sb.append('/'); -154sb.append(Bytes.toString(name)); -155sb.append("/multiget/"); -156if (rows == null || rows.length == 0) { -157 return sb.toString(); -158} -159sb.append("?"); -160for(int i=0; irows.length; i++) { -161 byte[] rk = rows[i]; -162 if (i != 0) { -163sb.append(''); -164 } -165 sb.append("row="); -166 sb.append(toURLEncodedBytes(rk)); -167} -168sb.append("v="); -169sb.append(maxVersions); -170 -171
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b3b50f22/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html index 249d4a0..7369fdf 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html @@ -65,12 +65,12 @@ 057import com.google.common.base.Preconditions; 058 059/** -060 * Reads {@link HFile} version 2 blocks to HFiles and via {@link Cacheable} Interface to caches. -061 * Version 2 was introduced in hbase-0.92.0. No longer has support for version 1 blocks since -062 * hbase-1.3.0. -063 * -064 * pVersion 1 was the original file block. Version 2 was introduced when we changed the hbase file -065 * format to support multi-level block indexes and compound bloom filters (HBASE-3857). +060 * Cacheable Blocks of an {@link HFile} version 2 file. +061 * Version 2 was introduced in hbase-0.92.0. +062 * +063 * pVersion 1 was the original file block. Version 2 was introduced when we changed the hbase file +064 * format to support multi-level block indexes and compound bloom filters (HBASE-3857). Support +065 * for Version 1 was removed in hbase-1.3.0. 066 * 067 * h3HFileBlock: Version 2/h3 068 * In version 2, a block is structured as follows: @@ -120,582 +120,582 @@ 112public class HFileBlock implements Cacheable { 113 private static final Log LOG = LogFactory.getLog(HFileBlock.class); 114 -115 /** Type of block. Header field 0. */ -116 private BlockType blockType; -117 -118 /** -119 * Size on disk excluding header, including checksum. Header field 1. -120 * @see Writer#putHeader(byte[], int, int, int, int) -121 */ -122 private int onDiskSizeWithoutHeader; -123 -124 /** -125 * Size of pure data. Does not include header or checksums. Header field 2. -126 * @see Writer#putHeader(byte[], int, int, int, int) -127 */ -128 private int uncompressedSizeWithoutHeader; -129 -130 /** -131 * The offset of the previous block on disk. Header field 3. -132 * @see Writer#putHeader(byte[], int, int, int, int) -133 */ -134 private long prevBlockOffset; -135 -136 /** -137 * Size on disk of header + data. Excludes checksum. Header field 6, -138 * OR calculated from {@link #onDiskSizeWithoutHeader} when using HDFS checksum. -139 * @see Writer#putHeader(byte[], int, int, int, int) -140 */ -141 private int onDiskDataSizeWithHeader; -142 -143 -144 /** -145 * The in-memory representation of the hfile block. Can be on or offheap. Can be backed by -146 * a single ByteBuffer or by many. Make no assumptions. -147 * -148 * pBe careful reading from this codebuf/code. Duplicate and work on the duplicate or if -149 * not, be sure to reset position and limit else trouble down the road. -150 * -151 * pTODO: Make this read-only once made. -152 * -153 * pWe are using the ByteBuff type. ByteBuffer is not extensible yet we need to be able to have -154 * a ByteBuffer-like API across multiple ByteBuffers reading from a cache such as BucketCache. -155 * So, we have this ByteBuff type. Unfortunately, it is spread all about HFileBlock. Would be -156 * good if could be confined to cache-use only but hard-to-do. -157 */ -158 private ByteBuff buf; -159 -160 /** Meta data that holds meta information on the hfileblock. -161 */ -162 private HFileContext fileContext; -163 -164 /** -165 * The offset of this block in the file. Populated by the reader for -166 * convenience of access. This offset is not part of the block header. -167 */ -168 private long offset = UNSET; -169 -170 private MemoryType memType = MemoryType.EXCLUSIVE; -171 -172 /** -173 * The on-disk size of the next block, including the header and checksums if present, obtained by -174 * peeking into the first {@link HConstants#HFILEBLOCK_HEADER_SIZE} bytes of the next block's -175 * header, or UNSET if unknown. -176 * -177 * Blocks try to carry the size of the next block to read in this data member. They will even have -178 * this value when served from cache. Could save a seek in the case where we are iterating through -179 * a file and some of the blocks come from cache. If from cache, then having this info to hand -180 * will save us doing a seek to read the header so we can read the body of a block. -181 * TODO: see how effective this is at saving seeks. -182 */ -183 private int nextBlockOnDiskSize = UNSET; -184 -185 /** -186 * On a checksum failure, do these many succeeding read requests using hdfs checksums before -187 * auto-reenabling hbase checksum verification. -188 */ -189 static final int
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ca9f6925/devapidocs/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.DisableTableProcedureBiConsumer.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.DisableTableProcedureBiConsumer.html b/devapidocs/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.DisableTableProcedureBiConsumer.html deleted file mode 100644 index 5e6de54..000 --- a/devapidocs/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.DisableTableProcedureBiConsumer.html +++ /dev/null @@ -1,339 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd;> - - - - - -AsyncHBaseAdmin.DisableTableProcedureBiConsumer (Apache HBase 3.0.0-SNAPSHOT API) - - - - - -var methods = {"i0":10}; -var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; -var altColor = "altColor"; -var rowColor = "rowColor"; -var tableTab = "tableTab"; -var activeTableTab = "activeTableTab"; - - -JavaScript is disabled on your browser. - - - - - -Skip navigation links - - - - -Overview -Package -Class -Use -Tree -Deprecated -Index -Help - - - - -PrevClass -NextClass - - -Frames -NoFrames - - -AllClasses - - - - - - - -Summary: -Nested| -Field| -Constr| -Method - - -Detail: -Field| -Constr| -Method - - - - - - - - -org.apache.hadoop.hbase.client -Class AsyncHBaseAdmin.DisableTableProcedureBiConsumer - - - -http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object - - -org.apache.hadoop.hbase.client.AsyncHBaseAdmin.ProcedureBiConsumer - - -org.apache.hadoop.hbase.client.AsyncHBaseAdmin.TableProcedureBiConsumer - - -org.apache.hadoop.hbase.client.AsyncHBaseAdmin.DisableTableProcedureBiConsumer - - - - - - - - - - - -All Implemented Interfaces: -http://docs.oracle.com/javase/8/docs/api/java/util/function/BiConsumer.html?is-external=true; title="class or interface in java.util.function">BiConsumerhttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void,http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwable - - -Enclosing class: -AsyncHBaseAdmin - - - -private class AsyncHBaseAdmin.DisableTableProcedureBiConsumer -extends AsyncHBaseAdmin.TableProcedureBiConsumer - - - - - - - - - - - -Field Summary - - - - -Fields inherited from classorg.apache.hadoop.hbase.client.AsyncHBaseAdmin.TableProcedureBiConsumer -tableName - - - - - -Fields inherited from classorg.apache.hadoop.hbase.client.AsyncHBaseAdmin.ProcedureBiConsumer -admin - - - - - - - - -Constructor Summary - -Constructors - -Constructor and Description - - -DisableTableProcedureBiConsumer(AsyncAdminadmin, - TableNametableName) - - - - - - - - - -Method Summary - -All MethodsInstance MethodsConcrete Methods - -Modifier and Type -Method and Description - - -(package private) http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String -getOperationType() - - - - - - -Methods inherited from classorg.apache.hadoop.hbase.client.AsyncHBaseAdmin.TableProcedureBiConsumer -getDescription, onError, onFinished - - - - - -Methods inherited from classorg.apache.hadoop.hbase.client.AsyncHBaseAdmin.ProcedureBiConsumer -accept - - - - - -Methods inherited from classjava.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object -http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--; title="class or interface in java.lang">clone, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-; title="class or interface in java.lang">equals, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--; title="class or interface in java.lang">finalize, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--; title="class or interface in java.lang">getClass, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--; title="class or interface in java.lang">hashCode, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--; title="class or interface in java.lang">notify,
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8e3b63ca/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.MasterRpcCall.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.MasterRpcCall.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.MasterRpcCall.html index dc12c09..82506d2 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.MasterRpcCall.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.MasterRpcCall.html @@ -54,2261 +54,2259 @@ 046import org.apache.commons.io.IOUtils; 047import org.apache.commons.logging.Log; 048import org.apache.commons.logging.LogFactory; -049import org.apache.directory.api.util.OptionalComponentsMonitor; -050import org.apache.hadoop.hbase.HRegionInfo; -051import org.apache.hadoop.hbase.HRegionLocation; -052import org.apache.hadoop.hbase.MetaTableAccessor; -053import org.apache.hadoop.hbase.MetaTableAccessor.QueryType; -054import org.apache.hadoop.hbase.NotServingRegionException; -055import org.apache.hadoop.hbase.ProcedureInfo; -056import org.apache.hadoop.hbase.RegionLocations; -057import org.apache.hadoop.hbase.ServerName; -058import org.apache.hadoop.hbase.NamespaceDescriptor; -059import org.apache.hadoop.hbase.HConstants; -060import org.apache.hadoop.hbase.TableExistsException; -061import org.apache.hadoop.hbase.TableName; -062import org.apache.hadoop.hbase.AsyncMetaTableAccessor; -063import org.apache.hadoop.hbase.TableNotDisabledException; -064import org.apache.hadoop.hbase.TableNotEnabledException; -065import org.apache.hadoop.hbase.TableNotFoundException; -066import org.apache.hadoop.hbase.UnknownRegionException; -067import org.apache.hadoop.hbase.classification.InterfaceAudience; -068import org.apache.hadoop.hbase.classification.InterfaceStability; -069import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder; -070import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder; -071import org.apache.hadoop.hbase.client.Scan.ReadType; -072import org.apache.hadoop.hbase.client.replication.ReplicationSerDeHelper; -073import org.apache.hadoop.hbase.client.replication.TableCFs; -074import org.apache.hadoop.hbase.exceptions.DeserializationException; -075import org.apache.hadoop.hbase.ipc.HBaseRpcController; -076import org.apache.hadoop.hbase.quotas.QuotaFilter; -077import org.apache.hadoop.hbase.quotas.QuotaSettings; -078import org.apache.hadoop.hbase.quotas.QuotaTableUtil; -079import org.apache.hadoop.hbase.replication.ReplicationException; -080import org.apache.hadoop.hbase.replication.ReplicationPeerConfig; -081import org.apache.hadoop.hbase.replication.ReplicationPeerDescription; -082import org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback; -083import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -084import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter; -085import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService; -086import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest; -087import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse; -088import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest; -089import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse; -090import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest; -091import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionResponse; -092import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionRequest; -093import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionResponse; -094import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionRequest; -095import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionResponse; -096import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription; -097import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema; -098import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest; -099import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse; -100import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest; -101import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse; -102import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest; -103import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse; -104import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest; -105import
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aecb1286/devapidocs/src-html/org/apache/hadoop/hbase/AsyncMetaTableAccessor.MetaTableRawScanResultConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/AsyncMetaTableAccessor.MetaTableRawScanResultConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/AsyncMetaTableAccessor.MetaTableRawScanResultConsumer.html index baee8c1..a0256ee 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/AsyncMetaTableAccessor.MetaTableRawScanResultConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/AsyncMetaTableAccessor.MetaTableRawScanResultConsumer.html @@ -30,551 +30,607 @@ 022import java.io.IOException; 023import java.util.ArrayList; 024import java.util.Arrays; -025import java.util.List; -026import java.util.Map; -027import java.util.NavigableMap; -028import java.util.Optional; -029import java.util.SortedMap; -030import java.util.concurrent.CompletableFuture; -031import java.util.regex.Matcher; -032import java.util.regex.Pattern; -033 -034import org.apache.commons.logging.Log; -035import org.apache.commons.logging.LogFactory; -036import org.apache.hadoop.hbase.MetaTableAccessor.CollectingVisitor; -037import org.apache.hadoop.hbase.MetaTableAccessor.QueryType; -038import org.apache.hadoop.hbase.MetaTableAccessor.Visitor; -039import org.apache.hadoop.hbase.classification.InterfaceAudience; -040import org.apache.hadoop.hbase.client.Consistency; -041import org.apache.hadoop.hbase.client.Get; -042import org.apache.hadoop.hbase.client.RawAsyncTable; -043import org.apache.hadoop.hbase.client.RawScanResultConsumer; -044import org.apache.hadoop.hbase.client.RegionReplicaUtil; -045import org.apache.hadoop.hbase.client.Result; -046import org.apache.hadoop.hbase.client.Scan; -047import org.apache.hadoop.hbase.client.TableState; -048import org.apache.hadoop.hbase.exceptions.DeserializationException; -049import org.apache.hadoop.hbase.util.Bytes; -050import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -051import org.apache.hadoop.hbase.util.Pair; -052 -053/** -054 * The asynchronous meta table accessor. Used to read/write region and assignment information store -055 * in codehbase:meta/code. -056 */ -057@InterfaceAudience.Private -058public class AsyncMetaTableAccessor { -059 -060 private static final Log LOG = LogFactory.getLog(AsyncMetaTableAccessor.class); -061 -062 -063 /** The delimiter for meta columns for replicaIds gt; 0 */ -064 private static final char META_REPLICA_ID_DELIMITER = '_'; +025import java.util.Collections; +026import java.util.List; +027import java.util.Map; +028import java.util.NavigableMap; +029import java.util.Optional; +030import java.util.SortedMap; +031import java.util.concurrent.CompletableFuture; +032import java.util.regex.Matcher; +033import java.util.regex.Pattern; +034import java.util.stream.Collectors; +035 +036import org.apache.commons.logging.Log; +037import org.apache.commons.logging.LogFactory; +038import org.apache.hadoop.hbase.MetaTableAccessor.CollectingVisitor; +039import org.apache.hadoop.hbase.MetaTableAccessor.QueryType; +040import org.apache.hadoop.hbase.MetaTableAccessor.Visitor; +041import org.apache.hadoop.hbase.classification.InterfaceAudience; +042import org.apache.hadoop.hbase.client.Connection; +043import org.apache.hadoop.hbase.client.Consistency; +044import org.apache.hadoop.hbase.client.Get; +045import org.apache.hadoop.hbase.client.RawAsyncTable; +046import org.apache.hadoop.hbase.client.RawScanResultConsumer; +047import org.apache.hadoop.hbase.client.RegionReplicaUtil; +048import org.apache.hadoop.hbase.client.Result; +049import org.apache.hadoop.hbase.client.Scan; +050import org.apache.hadoop.hbase.client.TableState; +051import org.apache.hadoop.hbase.client.Scan.ReadType; +052import org.apache.hadoop.hbase.exceptions.DeserializationException; +053import org.apache.hadoop.hbase.util.Bytes; +054import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; +055import org.apache.hadoop.hbase.util.Pair; +056 +057/** +058 * The asynchronous meta table accessor. Used to read/write region and assignment information store +059 * in codehbase:meta/code. +060 */ +061@InterfaceAudience.Private +062public class AsyncMetaTableAccessor { +063 +064 private static final Log LOG = LogFactory.getLog(AsyncMetaTableAccessor.class); 065 -066 /** A regex for parsing server columns from meta. See above javadoc for meta layout */ -067 private static final Pattern SERVER_COLUMN_PATTERN = Pattern -068 .compile("^server(_[0-9a-fA-F]{4})?$"); +066 +067 /** The delimiter for meta columns for replicaIds gt; 0 */ +068 private static final char META_REPLICA_ID_DELIMITER = '_'; 069 -070 public static CompletableFutureBoolean tableExists(RawAsyncTable metaTable, TableName tableName) { -071if (tableName.equals(META_TABLE_NAME)) { -072 return
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a719cd00/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html index 6de986f..c895448 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html @@ -26,1592 +26,1693 @@ 018package org.apache.hadoop.hbase.master.balancer; 019 020import java.util.ArrayDeque; -021import java.util.Arrays; -022import java.util.Collection; -023import java.util.Deque; -024import java.util.HashMap; -025import java.util.LinkedList; -026import java.util.List; -027import java.util.Map; -028import java.util.Map.Entry; -029import java.util.Random; -030 -031import org.apache.commons.logging.Log; -032import org.apache.commons.logging.LogFactory; -033import org.apache.hadoop.conf.Configuration; -034import org.apache.hadoop.hbase.ClusterStatus; -035import org.apache.hadoop.hbase.HBaseInterfaceAudience; -036import org.apache.hadoop.hbase.HConstants; -037import org.apache.hadoop.hbase.HRegionInfo; -038import org.apache.hadoop.hbase.RegionLoad; -039import org.apache.hadoop.hbase.ServerLoad; -040import org.apache.hadoop.hbase.ServerName; -041import org.apache.hadoop.hbase.TableName; -042import org.apache.hadoop.hbase.classification.InterfaceAudience; -043import org.apache.hadoop.hbase.master.MasterServices; -044import org.apache.hadoop.hbase.master.RegionPlan; -045import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action; -046import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action.Type; -047import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.AssignRegionAction; -048import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.MoveRegionAction; -049import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.SwapRegionsAction; -050import org.apache.hadoop.hbase.util.Bytes; -051import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -052 -053import com.google.common.collect.Lists; -054 -055/** -056 * pThis is a best effort load balancer. Given a Cost function F(C) =gt; x It will -057 * randomly try and mutate the cluster to Cprime. If F(Cprime) lt; F(C) then the -058 * new cluster state becomes the plan. It includes costs functions to compute the cost of:/p -059 * ul -060 * liRegion Load/li -061 * liTable Load/li -062 * liData Locality/li -063 * liMemstore Sizes/li -064 * liStorefile Sizes/li -065 * /ul -066 * -067 * -068 * pEvery cost function returns a number between 0 and 1 inclusive; where 0 is the lowest cost -069 * best solution, and 1 is the highest possible cost and the worst solution. The computed costs are -070 * scaled by their respective multipliers:/p +021import java.util.ArrayList; +022import java.util.Arrays; +023import java.util.Collection; +024import java.util.Collections; +025import java.util.Deque; +026import java.util.HashMap; +027import java.util.LinkedList; +028import java.util.List; +029import java.util.Map; +030import java.util.Map.Entry; +031import java.util.Random; +032 +033import org.apache.commons.logging.Log; +034import org.apache.commons.logging.LogFactory; +035import org.apache.hadoop.conf.Configuration; +036import org.apache.hadoop.hbase.ClusterStatus; +037import org.apache.hadoop.hbase.HBaseInterfaceAudience; +038import org.apache.hadoop.hbase.HConstants; +039import org.apache.hadoop.hbase.HRegionInfo; +040import org.apache.hadoop.hbase.RegionLoad; +041import org.apache.hadoop.hbase.ServerLoad; +042import org.apache.hadoop.hbase.ServerName; +043import org.apache.hadoop.hbase.TableName; +044import org.apache.hadoop.hbase.classification.InterfaceAudience; +045import org.apache.hadoop.hbase.master.MasterServices; +046import org.apache.hadoop.hbase.master.RegionPlan; +047import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action; +048import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action.Type; +049import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.AssignRegionAction; +050import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.LocalityType; +051import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.MoveRegionAction; +052import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.SwapRegionsAction; +053import org.apache.hadoop.hbase.util.Bytes; +054import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; +055 +056import com.google.common.base.Optional; +057import
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/476c54ed/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html index 0c3fe3b..d262744 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html @@ -149,3339 +149,3348 @@ 141import org.apache.hadoop.hbase.quotas.MasterSpaceQuotaObserver; 142import org.apache.hadoop.hbase.quotas.QuotaObserverChore; 143import org.apache.hadoop.hbase.quotas.QuotaUtil; -144import org.apache.hadoop.hbase.quotas.SpaceQuotaSnapshotNotifier; -145import org.apache.hadoop.hbase.quotas.SpaceQuotaSnapshotNotifierFactory; -146import org.apache.hadoop.hbase.regionserver.DefaultStoreEngine; -147import org.apache.hadoop.hbase.regionserver.HRegionServer; -148import org.apache.hadoop.hbase.regionserver.HStore; -149import org.apache.hadoop.hbase.regionserver.RSRpcServices; -150import org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost; -151import org.apache.hadoop.hbase.regionserver.RegionSplitPolicy; -152import org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy; -153import org.apache.hadoop.hbase.regionserver.compactions.FIFOCompactionPolicy; -154import org.apache.hadoop.hbase.replication.ReplicationException; -155import org.apache.hadoop.hbase.replication.ReplicationFactory; -156import org.apache.hadoop.hbase.replication.ReplicationPeerConfig; -157import org.apache.hadoop.hbase.replication.ReplicationPeerDescription; -158import org.apache.hadoop.hbase.replication.ReplicationQueuesZKImpl; -159import org.apache.hadoop.hbase.replication.master.TableCFsUpdater; -160import org.apache.hadoop.hbase.replication.regionserver.Replication; -161import org.apache.hadoop.hbase.security.AccessDeniedException; -162import org.apache.hadoop.hbase.security.UserProvider; -163import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -164import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState; -165import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionServerInfo; -166import org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas; -167import org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceViolationPolicy; -168import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription; -169import org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos; -170import org.apache.hadoop.hbase.util.Addressing; -171import org.apache.hadoop.hbase.util.Bytes; -172import org.apache.hadoop.hbase.util.CompressionTest; -173import org.apache.hadoop.hbase.util.EncryptionTest; -174import org.apache.hadoop.hbase.util.FSUtils; -175import org.apache.hadoop.hbase.util.HFileArchiveUtil; -176import org.apache.hadoop.hbase.util.HasThread; -177import org.apache.hadoop.hbase.util.IdLock; -178import org.apache.hadoop.hbase.util.ModifyRegionUtils; -179import org.apache.hadoop.hbase.util.Pair; -180import org.apache.hadoop.hbase.util.Threads; -181import org.apache.hadoop.hbase.util.VersionInfo; -182import org.apache.hadoop.hbase.util.ZKDataMigrator; -183import org.apache.hadoop.hbase.zookeeper.DrainingServerTracker; -184import org.apache.hadoop.hbase.zookeeper.LoadBalancerTracker; -185import org.apache.hadoop.hbase.zookeeper.MasterAddressTracker; -186import org.apache.hadoop.hbase.zookeeper.MasterMaintenanceModeTracker; -187import org.apache.hadoop.hbase.zookeeper.RegionNormalizerTracker; -188import org.apache.hadoop.hbase.zookeeper.RegionServerTracker; -189import org.apache.hadoop.hbase.zookeeper.SplitOrMergeTracker; -190import org.apache.hadoop.hbase.zookeeper.ZKClusterId; -191import org.apache.hadoop.hbase.zookeeper.ZKUtil; -192import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; -193import org.apache.zookeeper.KeeperException; -194import org.eclipse.jetty.server.Server; -195import org.eclipse.jetty.server.ServerConnector; -196import org.eclipse.jetty.servlet.ServletHolder; -197import org.eclipse.jetty.webapp.WebAppContext; -198 -199import com.google.common.annotations.VisibleForTesting; -200import com.google.common.collect.Lists; -201import com.google.common.collect.Maps; -202import com.google.protobuf.Descriptors; -203import com.google.protobuf.Service; -204 -205/** -206 * HMaster is the "master server" for HBase. An HBase cluster has one active -207 * master. If many masters are started, all compete. Whichever wins goes on to -208 * run the cluster. All others park themselves in their constructor until -209 * master or cluster shutdown or until the active master loses its lease in -210 * zookeeper. Thereafter, all
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/77a552c4/devapidocs/org/apache/hadoop/hbase/HTableDescriptor.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/HTableDescriptor.html b/devapidocs/org/apache/hadoop/hbase/HTableDescriptor.html index bb8fd51..6831fb2 100644 --- a/devapidocs/org/apache/hadoop/hbase/HTableDescriptor.html +++ b/devapidocs/org/apache/hadoop/hbase/HTableDescriptor.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":42,"i1":42,"i2":42,"i3":42,"i4":42,"i5":42,"i6":42,"i7":42,"i8":42,"i9":42,"i10":42,"i11":42,"i12":42,"i13":42,"i14":42,"i15":42,"i16":42,"i17":42,"i18":42,"i19":42,"i20":42,"i21":42,"i22":42,"i23":42,"i24":42,"i25":42,"i26":42,"i27":42,"i28":42,"i29":42,"i30":42,"i31":42,"i32":42,"i33":42,"i34":42,"i35":42,"i36":42,"i37":42,"i38":42,"i39":41,"i40":42,"i41":42,"i42":42,"i43":42,"i44":42,"i45":42,"i46":42,"i47":42,"i48":42,"i49":42,"i50":42,"i51":42,"i52":42,"i53":42,"i54":42,"i55":42,"i56":42,"i57":42,"i58":42,"i59":42,"i60":42,"i61":42,"i62":42,"i63":42,"i64":42,"i65":42,"i66":42}; +var methods = {"i0":42,"i1":42,"i2":42,"i3":42,"i4":42,"i5":42,"i6":42,"i7":42,"i8":42,"i9":42,"i10":42,"i11":42,"i12":42,"i13":42,"i14":42,"i15":42,"i16":42,"i17":42,"i18":42,"i19":42,"i20":42,"i21":42,"i22":42,"i23":42,"i24":42,"i25":42,"i26":42,"i27":42,"i28":42,"i29":42,"i30":42,"i31":42,"i32":42,"i33":42,"i34":42,"i35":42,"i36":42,"i37":42,"i38":42,"i39":42,"i40":42,"i41":42,"i42":42,"i43":42,"i44":41,"i45":42,"i46":42,"i47":42,"i48":42,"i49":42,"i50":42,"i51":42,"i52":42,"i53":42,"i54":42,"i55":42,"i56":42,"i57":42,"i58":42,"i59":42,"i60":42,"i61":42,"i62":42,"i63":42,"i64":42,"i65":42,"i66":42,"i67":42,"i68":42,"i69":42,"i70":42,"i71":42,"i72":42}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"],32:["t6","Deprecated Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -123,7 +123,7 @@ var activeTableTab = "activeTableTab"; http://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true; title="class or interface in java.lang">@Deprecated @InterfaceAudience.Public -public class HTableDescriptor +public class HTableDescriptor extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements TableDescriptor, http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableHTableDescriptor HTableDescriptor contains the details about an HBase table such as the descriptors of @@ -191,7 +191,7 @@ implements -private TableDescriptorBuilder.ModifyableTableDescriptor +protected TableDescriptorBuilder.ModifyableTableDescriptor delegatee Deprecated. @@ -306,6 +306,13 @@ implements + + + +Fields inherited from interfaceorg.apache.hadoop.hbase.client.TableDescriptor +COMPARATOR + @@ -328,33 +335,31 @@ implements +protected +HTableDescriptor(HTableDescriptordesc, +booleandeepClone) +Deprecated. + + + HTableDescriptor(TableDescriptordesc) Deprecated. - + protected HTableDescriptor(TableDescriptorBuilder.ModifyableTableDescriptordelegatee) Deprecated. - + HTableDescriptor(TableNamename) Deprecated. Construct a table descriptor specifying a TableName object - -private -HTableDescriptor(TableNamename, -http://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true; title="class or interface in java.util">CollectionHColumnDescriptorfamilies, -http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapBytes,Bytesvalues, -http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringconfiguration) -Deprecated. - - HTableDescriptor(TableNamename, @@ -428,69 +433,89 @@ implements HColumnDescriptor[] getColumnFamilies() Deprecated. -Returns an array all the HColumnDescriptor of the column families - of the table. +ColumnFamilyDescriptor +getColumnFamily(byte[]name) +Deprecated. +Returns the ColumnDescriptor for a specific column family with name as + specified by the parameter column. + + + int getColumnFamilyCount() Deprecated. Returns the count of the column families of the table. - + +http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">Setbyte[] +getColumnFamilyNames() +Deprecated. +Returns all the column family names of the current table.
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b44796ef/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html b/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html index b7c90be..f26f8aa 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html @@ -193,992 +193,1035 @@ 185 * Control whether to use pread at server side. 186 */ 187 private ReadType readType = ReadType.DEFAULT; -188 /** -189 * Create a Scan operation across all rows. -190 */ -191 public Scan() {} -192 -193 /** -194 * @deprecated use {@code new Scan().withStartRow(startRow).setFilter(filter)} instead. -195 */ -196 @Deprecated -197 public Scan(byte[] startRow, Filter filter) { -198this(startRow); -199this.filter = filter; -200 } -201 -202 /** -203 * Create a Scan operation starting at the specified row. -204 * p -205 * If the specified row does not exist, the Scanner will start from the next closest row after the -206 * specified row. -207 * @param startRow row to start scanner at or after -208 * @deprecated use {@code new Scan().withStartRow(startRow)} instead. -209 */ -210 @Deprecated -211 public Scan(byte[] startRow) { -212setStartRow(startRow); -213 } -214 -215 /** -216 * Create a Scan operation for the range of rows specified. -217 * @param startRow row to start scanner at or after (inclusive) -218 * @param stopRow row to stop scanner before (exclusive) -219 * @deprecated use {@code new Scan().withStartRow(startRow).withStopRow(stopRow)} instead. -220 */ -221 @Deprecated -222 public Scan(byte[] startRow, byte[] stopRow) { -223setStartRow(startRow); -224setStopRow(stopRow); -225 } -226 -227 /** -228 * Creates a new instance of this class while copying all values. -229 * -230 * @param scan The scan instance to copy from. -231 * @throws IOException When copying the values fails. -232 */ -233 public Scan(Scan scan) throws IOException { -234startRow = scan.getStartRow(); -235includeStartRow = scan.includeStartRow(); -236stopRow = scan.getStopRow(); -237includeStopRow = scan.includeStopRow(); -238maxVersions = scan.getMaxVersions(); -239batch = scan.getBatch(); -240storeLimit = scan.getMaxResultsPerColumnFamily(); -241storeOffset = scan.getRowOffsetPerColumnFamily(); -242caching = scan.getCaching(); -243maxResultSize = scan.getMaxResultSize(); -244cacheBlocks = scan.getCacheBlocks(); -245filter = scan.getFilter(); // clone? -246loadColumnFamiliesOnDemand = scan.getLoadColumnFamiliesOnDemandValue(); -247consistency = scan.getConsistency(); -248 this.setIsolationLevel(scan.getIsolationLevel()); -249reversed = scan.isReversed(); -250asyncPrefetch = scan.isAsyncPrefetch(); -251small = scan.isSmall(); -252allowPartialResults = scan.getAllowPartialResults(); -253tr = scan.getTimeRange(); // TimeRange is immutable -254Mapbyte[], NavigableSetbyte[] fams = scan.getFamilyMap(); -255for (Map.Entrybyte[],NavigableSetbyte[] entry : fams.entrySet()) { -256 byte [] fam = entry.getKey(); -257 NavigableSetbyte[] cols = entry.getValue(); -258 if (cols != null cols.size() 0) { -259for (byte[] col : cols) { -260 addColumn(fam, col); -261} -262 } else { -263addFamily(fam); -264 } -265} -266for (Map.EntryString, byte[] attr : scan.getAttributesMap().entrySet()) { -267 setAttribute(attr.getKey(), attr.getValue()); +188 +189 private boolean needCursorResult = false; +190 +191 /** +192 * Create a Scan operation across all rows. +193 */ +194 public Scan() {} +195 +196 /** +197 * @deprecated use {@code new Scan().withStartRow(startRow).setFilter(filter)} instead. +198 */ +199 @Deprecated +200 public Scan(byte[] startRow, Filter filter) { +201this(startRow); +202this.filter = filter; +203 } +204 +205 /** +206 * Create a Scan operation starting at the specified row. +207 * p +208 * If the specified row does not exist, the Scanner will start from the next closest row after the +209 * specified row. +210 * @param startRow row to start scanner at or after +211 * @deprecated use {@code new Scan().withStartRow(startRow)} instead. +212 */ +213 @Deprecated +214 public Scan(byte[] startRow) { +215setStartRow(startRow); +216 } +217 +218 /** +219 * Create a Scan operation for the range of rows specified. +220 * @param startRow row to start scanner at or after (inclusive) +221 * @param stopRow row to stop scanner before (exclusive) +222 * @deprecated use {@code new Scan().withStartRow(startRow).withStopRow(stopRow)} instead. +223 */ +224 @Deprecated +225 public Scan(byte[] startRow, byte[] stopRow)
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6ad4f21a/devapidocs/org/apache/hadoop/hbase/master/procedure/DisableTableProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/DisableTableProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/DisableTableProcedure.html index d98b54f..1c410ff 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/DisableTableProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/DisableTableProcedure.html @@ -389,7 +389,7 @@ extends Procedure -addStackIndex, afterReplay, beforeReplay, compareTo, completionCleanup, doAcquireLock, doExecute, d oReleaseLock, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcId, getProcIdHashCode, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, < a href="../../../../../../org/apache/hadoop/hbase/procedure2/Procedure.html#getTimeoutTimestamp--">getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, removeStackIndex, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, shouldWaitClientAck, toString, toStringClass, toStringDetails, toStringSimpleSB, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, wasExecuted +addStackIndex, afterReplay, beforeReplay, compareTo, completionCleanup, doAcquireLock, doExecute, d oReleaseLock, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, removeStackIndex, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, shouldWaitClientAck, toString, toStringClass, toStringDetails, toStringSimpleSB, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, wasExecuted http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6ad4f21a/devapidocs/org/apache/hadoop/hbase/master/procedure/DispatchMergingRegionsProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/DispatchMergingRegionsProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/DispatchMergingRegionsProcedure.html index 5708c32..6734a0e 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/DispatchMergingRegionsProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/DispatchMergingRegionsProcedure.html @@ -436,7 +436,7 @@ extends Procedure -addStackIndex, afterReplay, beforeReplay, compareTo, completionCleanup, doAcquireLock, doExecute, d oReleaseLock, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcId, getProcIdHashCode, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, < a href="../../../../../../org/apache/hadoop/hbase/procedure2/Procedure.html#getTimeoutTimestamp--">getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, removeStackIndex, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, shouldWaitClientAck, toString, toStringClass, toStringDetails, toStringSimpleSB, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, wasExecuted +addStackIndex, afterReplay, beforeReplay, compareTo, completionCleanup, doAcquireLock, doExecute, d oReleaseLock, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, getResult,
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9d35424/devapidocs/src-html/org/apache/hadoop/hbase/io/MetricsIOSource.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/io/MetricsIOSource.html b/devapidocs/src-html/org/apache/hadoop/hbase/io/MetricsIOSource.html new file mode 100644 index 000..b074547 --- /dev/null +++ b/devapidocs/src-html/org/apache/hadoop/hbase/io/MetricsIOSource.html @@ -0,0 +1,152 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + +Source code + + + + +001/** +002 * Licensed to the Apache Software Foundation (ASF) under one +003 * or more contributor license agreements. See the NOTICE file +004 * distributed with this work for additional information +005 * regarding copyright ownership. The ASF licenses this file +006 * to you under the Apache License, Version 2.0 (the +007 * "License"); you may not use this file except in compliance +008 * with the License. You may obtain a copy of the License at +009 * +010 * http://www.apache.org/licenses/LICENSE-2.0 +011 * +012 * Unless required by applicable law or agreed to in writing, software +013 * distributed under the License is distributed on an "AS IS" BASIS, +014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +015 * See the License for the specific language governing permissions and +016 * limitations under the License. +017 */ +018 +019package org.apache.hadoop.hbase.io; +020 +021import org.apache.hadoop.hbase.metrics.BaseSource; +022 +023public interface MetricsIOSource extends BaseSource { +024 +025 /** +026 * The name of the metrics +027 */ +028 String METRICS_NAME = "IO"; +029 +030 /** +031 * The name of the metrics context that metrics will be under. +032 */ +033 String METRICS_CONTEXT = "regionserver"; +034 +035 /** +036 * Description +037 */ +038 String METRICS_DESCRIPTION = "Metrics about FileSystem IO"; +039 +040 /** +041 * The name of the metrics context that metrics will be under in jmx +042 */ +043 String METRICS_JMX_CONTEXT = "RegionServer,sub=" + METRICS_NAME; +044 +045 +046 String FS_READ_TIME_HISTO_KEY = "fsReadTime"; +047 String FS_PREAD_TIME_HISTO_KEY = "fsPReadTime"; +048 String FS_WRITE_HISTO_KEY = "fsWriteTime"; +049 +050 String CHECKSUM_FAILURES_KEY = "fsChecksumFailureCount"; +051 +052 String FS_READ_TIME_HISTO_DESC +053= "Latency of HFile's sequential reads on this region server in milliseconds"; +054 String FS_PREAD_TIME_HISTO_DESC +055= "Latency of HFile's positional reads on this region server in milliseconds"; +056 String FS_WRITE_TIME_HISTO_DESC +057= "Latency of HFile's writes on this region server in milliseconds"; +058 +059 String CHECKSUM_FAILURES_DESC = "Number of checksum failures for the HBase HFile checksums at the" +060 + " HBase level (separate from HDFS checksums)"; +061 +062 +063 /** +064 * Update the fs sequential read time histogram +065 * @param t time it took, in milliseconds +066 */ +067 void updateFsReadTime(long t); +068 +069 /** +070 * Update the fs positional read time histogram +071 * @param t time it took, in milliseconds +072 */ +073 void updateFsPReadTime(long t); +074 +075 /** +076 * Update the fs write time histogram +077 * @param t time it took, in milliseconds +078 */ +079 void updateFsWriteTime(long t); +080} + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9d35424/devapidocs/src-html/org/apache/hadoop/hbase/io/MetricsIOSourceImpl.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/io/MetricsIOSourceImpl.html b/devapidocs/src-html/org/apache/hadoop/hbase/io/MetricsIOSourceImpl.html new file mode 100644 index 000..11c20e6 --- /dev/null +++ b/devapidocs/src-html/org/apache/hadoop/hbase/io/MetricsIOSourceImpl.html @@ -0,0 +1,156 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + +Source code + + + + +001/** +002 * Licensed to the Apache Software Foundation (ASF) under one +003 * or more contributor license agreements. See the NOTICE file +004 * distributed with this work for additional information +005 * regarding copyright ownership. The ASF licenses this file +006 * to you under the Apache License, Version 2.0 (the +007 * "License"); you may not use this file except in compliance +008 * with the License. You may obtain a copy of the License at +009 * +010 * http://www.apache.org/licenses/LICENSE-2.0 +011 * +012 * Unless required by applicable law or agreed to in writing, software +013 * distributed under the License is distributed on an "AS IS" BASIS, +014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +015 * See the License for the specific language governing permissions and +016 *
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6dd31117/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html b/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html index 025ba8d..25b805a 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html +++ b/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -243,301 +243,311 @@ implements +org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse +dispatchMergingRegions(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequestrequest) + + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse drainRegionServers(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequestrequest) - + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse enableCatalogJanitor(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequestrequest) - + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse enableReplicationPeer(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequestrequest) - + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse enableTable(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequestrequest) - + org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse execMasterService(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequestrequest) - + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse execProcedure(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequestrequest) - + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse execProcedureWithRet(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequestrequest) - + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse getClusterStatus(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequestrequest) - +
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dab57116/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.html b/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.html index a6ce1bc..c98e496 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.html +++ b/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.html @@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -class ConnectionImplementation +class ConnectionImplementation extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements ClusterConnection, http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true; title="class or interface in java.io">Closeable Main implementation of Connection and ClusterConnection interfaces. @@ -857,7 +857,7 @@ implements RETRIES_BY_SERVER_KEY -public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String RETRIES_BY_SERVER_KEY +public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String RETRIES_BY_SERVER_KEY See Also: Constant Field Values @@ -870,7 +870,7 @@ implements LOG -private static finalorg.apache.commons.logging.Log LOG +private static finalorg.apache.commons.logging.Log LOG @@ -879,7 +879,7 @@ implements RESOLVE_HOSTNAME_ON_FAIL_KEY -private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String RESOLVE_HOSTNAME_ON_FAIL_KEY +private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String RESOLVE_HOSTNAME_ON_FAIL_KEY See Also: Constant Field Values @@ -892,7 +892,7 @@ implements hostnamesCanChange -private finalboolean hostnamesCanChange +private finalboolean hostnamesCanChange @@ -901,7 +901,7 @@ implements pause -private finallong pause +private finallong pause @@ -910,7 +910,7 @@ implements pauseForCQTBE -private finallong pauseForCQTBE +private finallong pauseForCQTBE @@ -919,7 +919,7 @@ implements useMetaReplicas -privateboolean useMetaReplicas +privateboolean useMetaReplicas @@ -928,7 +928,7 @@ implements metaReplicaCallTimeoutScanInMicroSecond -private finalint metaReplicaCallTimeoutScanInMicroSecond +private finalint metaReplicaCallTimeoutScanInMicroSecond @@ -937,7 +937,7 @@ implements numTries -private finalint numTries +private finalint numTries @@ -946,7 +946,7 @@ implements rpcTimeout -finalint rpcTimeout +finalint rpcTimeout @@ -955,7 +955,7 @@ implements nonceGenerator -private static volatileNonceGenerator nonceGenerator +private static volatileNonceGenerator nonceGenerator Global nonceGenerator shared per client.Currently there's no reason to limit its scope. Once it's set under nonceGeneratorCreateLock, it is never unset or changed. @@ -966,7 +966,7 @@ implements nonceGeneratorCreateLock -private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object nonceGeneratorCreateLock +private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object nonceGeneratorCreateLock The nonce generator lock. Only taken when creating Connection, which gets a private copy. @@ -976,7 +976,7 @@ implements asyncProcess -private finalAsyncProcess asyncProcess +private finalAsyncProcess asyncProcess @@ -985,7 +985,7 @@ implements stats -private finalServerStatisticTracker stats +private finalServerStatisticTracker stats @@ -994,7 +994,7 @@ implements closed -private volatileboolean closed +private volatileboolean closed @@ -1003,7 +1003,7 @@ implements aborted -private volatileboolean aborted +private volatileboolean aborted @@ -1012,7 +1012,7 @@ implements clusterStatusListener -ClusterStatusListener clusterStatusListener +ClusterStatusListener clusterStatusListener @@ -1021,7 +1021,7 @@ implements metaRegionLock -private finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object metaRegionLock +private finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object metaRegionLock @@ -1030,7 +1030,7 @@ implements masterAndZKLock -private
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c635e71b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html index be7f8e5..37574d7 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html @@ -147,1885 +147,1897 @@ 139 private final boolean hostnamesCanChange; 140 private final long pause; 141 private final long pauseForCQTBE;// pause for CallQueueTooBigException, if specified -142 private final boolean useMetaReplicas; -143 private final int numTries; -144 final int rpcTimeout; -145 -146 /** -147 * Global nonceGenerator shared per client.Currently there's no reason to limit its scope. -148 * Once it's set under nonceGeneratorCreateLock, it is never unset or changed. -149 */ -150 private static volatile NonceGenerator nonceGenerator = null; -151 /** The nonce generator lock. Only taken when creating Connection, which gets a private copy. */ -152 private static final Object nonceGeneratorCreateLock = new Object(); -153 -154 private final AsyncProcess asyncProcess; -155 // single tracker per connection -156 private final ServerStatisticTracker stats; -157 -158 private volatile boolean closed; -159 private volatile boolean aborted; -160 -161 // package protected for the tests -162 ClusterStatusListener clusterStatusListener; -163 -164 private final Object metaRegionLock = new Object(); -165 -166 // We have a single lock for master zk to prevent deadlocks. Having -167 // one lock for ZK and one lock for master is not possible: -168 // When creating a connection to master, we need a connection to ZK to get -169 // its address. But another thread could have taken the ZK lock, and could -170 // be waiting for the master lock = deadlock. -171 private final Object masterAndZKLock = new Object(); -172 -173 // thread executor shared by all Table instances created -174 // by this connection -175 private volatile ExecutorService batchPool = null; -176 // meta thread executor shared by all Table instances created -177 // by this connection -178 private volatile ExecutorService metaLookupPool = null; -179 private volatile boolean cleanupPool = false; -180 -181 private final Configuration conf; -182 -183 // cache the configuration value for tables so that we can avoid calling -184 // the expensive Configuration to fetch the value multiple times. -185 private final ConnectionConfiguration connectionConfig; -186 -187 // Client rpc instance. -188 private final RpcClient rpcClient; -189 -190 private final MetaCache metaCache; -191 private final MetricsConnection metrics; -192 -193 protected User user; -194 -195 private final RpcRetryingCallerFactory rpcCallerFactory; -196 -197 private final RpcControllerFactory rpcControllerFactory; -198 -199 private final RetryingCallerInterceptor interceptor; -200 -201 /** -202 * Cluster registry of basic info such as clusterid and meta region location. -203 */ -204 Registry registry; -205 -206 private final ClientBackoffPolicy backoffPolicy; -207 -208 /** -209 * Allow setting an alternate BufferedMutator implementation via -210 * config. If null, use default. -211 */ -212 private final String alternateBufferedMutatorClassName; -213 -214 /** -215 * constructor -216 * @param conf Configuration object -217 */ -218 ConnectionImplementation(Configuration conf, -219 ExecutorService pool, User user) throws IOException { -220this.conf = conf; -221this.user = user; -222this.batchPool = pool; -223this.connectionConfig = new ConnectionConfiguration(conf); -224this.closed = false; -225this.pause = conf.getLong(HConstants.HBASE_CLIENT_PAUSE, -226 HConstants.DEFAULT_HBASE_CLIENT_PAUSE); -227long configuredPauseForCQTBE = conf.getLong(HConstants.HBASE_CLIENT_PAUSE_FOR_CQTBE, pause); -228if (configuredPauseForCQTBE pause) { -229 LOG.warn("The " + HConstants.HBASE_CLIENT_PAUSE_FOR_CQTBE + " setting: " -230 + configuredPauseForCQTBE + " is smaller than " + HConstants.HBASE_CLIENT_PAUSE -231 + ", will use " + pause + " instead."); -232 this.pauseForCQTBE = pause; -233} else { -234 this.pauseForCQTBE = configuredPauseForCQTBE; -235} -236this.useMetaReplicas = conf.getBoolean(HConstants.USE_META_REPLICAS, -237 HConstants.DEFAULT_USE_META_REPLICAS); -238// how many times to try, one more than max *retry*
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8b4cf63f/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html index f81384b..f1c87a7 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html @@ -1192,6 +1192,14 @@ private void +CloneSnapshotProcedure.restoreSnapshotAcl(MasterProcedureEnvenv) + + +private void +RestoreSnapshotProcedure.restoreSnapshotAcl(MasterProcedureEnvenv) + + +private void ModifyColumnFamilyProcedure.restoreTableDescriptor(MasterProcedureEnvenv) Restore back to the old descriptor @@ -1637,59 +1645,65 @@ http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListHRegionInforegions) -CloneSnapshotProcedure(MasterProcedureEnvenv, +CloneSnapshotProcedure(MasterProcedureEnvenv, HTableDescriptorhTableDescriptor, - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.SnapshotDescriptionsnapshot) + org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionsnapshot) + + +CloneSnapshotProcedure(MasterProcedureEnvenv, + HTableDescriptorhTableDescriptor, + org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionsnapshot, + booleanrestoreAcl) Constructor - + CreateNamespaceProcedure(MasterProcedureEnvenv, NamespaceDescriptornsDescriptor) - + CreateTableProcedure(MasterProcedureEnvenv, HTableDescriptorhTableDescriptor, HRegionInfo[]newRegions) - + CreateTableProcedure(MasterProcedureEnvenv, HTableDescriptorhTableDescriptor, HRegionInfo[]newRegions, ProcedurePrepareLatchsyncLatch) - + DeleteColumnFamilyProcedure(MasterProcedureEnvenv, TableNametableName, byte[]familyName) - + DeleteColumnFamilyProcedure(MasterProcedureEnvenv, TableNametableName, byte[]familyName, ProcedurePrepareLatchlatch) - + DeleteNamespaceProcedure(MasterProcedureEnvenv, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringnamespaceName) - + DeleteTableProcedure(MasterProcedureEnvenv, TableNametableName) - + DeleteTableProcedure(MasterProcedureEnvenv, TableNametableName, ProcedurePrepareLatchsyncLatch) - + DisableTableProcedure(MasterProcedureEnvenv, TableNametableName, booleanskipTableStateCheck) Constructor - + DisableTableProcedure(MasterProcedureEnvenv, TableNametableName, booleanskipTableStateCheck, @@ -1697,14 +1711,14 @@ Constructor - + EnableTableProcedure(MasterProcedureEnvenv, TableNametableName, booleanskipTableStateCheck) Constructor - + EnableTableProcedure(MasterProcedureEnvenv, TableNametableName, booleanskipTableStateCheck, @@ -1712,39 +1726,45 @@ Constructor - + MergeTableRegionsProcedure(MasterProcedureEnvenv, HRegionInfo[]regionsToMerge, booleanforcible) - + ModifyColumnFamilyProcedure(MasterProcedureEnvenv, TableNametableName, HColumnDescriptorcfDescriptor) - + ModifyColumnFamilyProcedure(MasterProcedureEnvenv, TableNametableName, HColumnDescriptorcfDescriptor, ProcedurePrepareLatchlatch) - + ModifyNamespaceProcedure(MasterProcedureEnvenv, NamespaceDescriptornewNsDescriptor) - + ModifyTableProcedure(MasterProcedureEnvenv, HTableDescriptorhtd) - + ModifyTableProcedure(MasterProcedureEnvenv, HTableDescriptorhtd, ProcedurePrepareLatchlatch) + +RestoreSnapshotProcedure(MasterProcedureEnvenv, +HTableDescriptorhTableDescriptor, + org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionsnapshot) + -RestoreSnapshotProcedure(MasterProcedureEnvenv,
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html index 109b5f3..e484176 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html @@ -23,484 +23,310 @@ 015 * See the License for the specific language governing permissions and 016 * limitations under the License. 017 */ -018 -019package org.apache.hadoop.hbase.ipc; -020 -021import io.netty.bootstrap.ServerBootstrap; -022import io.netty.buffer.ByteBuf; -023import io.netty.buffer.PooledByteBufAllocator; -024import io.netty.buffer.Unpooled; -025import io.netty.channel.Channel; -026import io.netty.channel.ChannelFuture; -027import io.netty.channel.ChannelFutureListener; -028import io.netty.channel.ChannelHandlerContext; -029import io.netty.channel.ChannelInboundHandlerAdapter; -030import io.netty.channel.ChannelInitializer; -031import io.netty.channel.ChannelOption; -032import io.netty.channel.ChannelOutboundHandlerAdapter; -033import io.netty.channel.ChannelPipeline; -034import io.netty.channel.ChannelPromise; -035import io.netty.channel.EventLoopGroup; -036import io.netty.channel.epoll.EpollEventLoopGroup; -037import io.netty.channel.epoll.EpollServerSocketChannel; -038import io.netty.channel.group.ChannelGroup; -039import io.netty.channel.group.DefaultChannelGroup; -040import io.netty.channel.nio.NioEventLoopGroup; -041import io.netty.channel.socket.SocketChannel; -042import io.netty.channel.socket.nio.NioServerSocketChannel; -043import io.netty.handler.codec.ByteToMessageDecoder; -044import io.netty.handler.codec.LengthFieldBasedFrameDecoder; -045import io.netty.util.concurrent.GlobalEventExecutor; -046 -047import java.io.IOException; -048import java.io.InterruptedIOException; -049import java.net.InetAddress; -050import java.net.InetSocketAddress; -051import java.nio.ByteBuffer; -052import java.util.Arrays; -053import java.util.List; -054import java.util.concurrent.CountDownLatch; -055 -056import org.apache.commons.logging.Log; -057import org.apache.commons.logging.LogFactory; -058import org.apache.hadoop.conf.Configuration; -059import org.apache.hadoop.hbase.CellScanner; -060import org.apache.hadoop.hbase.HConstants; -061import org.apache.hadoop.hbase.Server; -062import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -063import org.apache.hadoop.hbase.nio.ByteBuff; -064import org.apache.hadoop.hbase.nio.SingleByteBuff; -065import org.apache.hadoop.hbase.security.AccessDeniedException; -066import org.apache.hadoop.hbase.security.AuthMethod; -067import org.apache.hadoop.hbase.security.HBasePolicyProvider; -068import org.apache.hadoop.hbase.security.SaslStatus; -069import org.apache.hadoop.hbase.security.SaslUtil; -070import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -071import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -072import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -073import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -074import org.apache.hadoop.hbase.util.Bytes; -075import org.apache.hadoop.hbase.util.JVM; -076import org.apache.hadoop.hbase.util.Pair; -077import org.apache.hadoop.io.IntWritable; -078import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -079import org.apache.htrace.TraceInfo; +018package org.apache.hadoop.hbase.ipc; +019 +020import io.netty.bootstrap.ServerBootstrap; +021import io.netty.buffer.ByteBuf; +022import io.netty.buffer.PooledByteBufAllocator; +023import io.netty.buffer.Unpooled; +024import io.netty.channel.Channel; +025import io.netty.channel.ChannelFuture; +026import io.netty.channel.ChannelFutureListener; +027import io.netty.channel.ChannelHandlerContext; +028import io.netty.channel.ChannelInboundHandlerAdapter; +029import io.netty.channel.ChannelInitializer; +030import io.netty.channel.ChannelOption; +031import io.netty.channel.ChannelOutboundHandlerAdapter; +032import io.netty.channel.ChannelPipeline; +033import io.netty.channel.ChannelPromise; +034import io.netty.channel.EventLoopGroup; +035import io.netty.channel.epoll.EpollEventLoopGroup; +036import io.netty.channel.epoll.EpollServerSocketChannel; +037import io.netty.channel.group.ChannelGroup; +038import io.netty.channel.group.DefaultChannelGroup; +039import io.netty.channel.nio.NioEventLoopGroup; +040import io.netty.channel.socket.SocketChannel; +041import io.netty.channel.socket.nio.NioServerSocketChannel; +042import io.netty.handler.codec.ByteToMessageDecoder; +043import
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8e0a5167/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html index cc40d21..2df0b04 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html @@ -61,8 +61,8 @@ 053import org.apache.hadoop.hbase.backup.BackupRestoreConstants; 054import org.apache.hadoop.hbase.backup.BackupRestoreConstants.BackupCommand; 055import org.apache.hadoop.hbase.backup.BackupType; -056import org.apache.hadoop.hbase.backup.util.BackupUtils; -057import org.apache.hadoop.hbase.backup.util.BackupSet; +056import org.apache.hadoop.hbase.backup.util.BackupSet; +057import org.apache.hadoop.hbase.backup.util.BackupUtils; 058import org.apache.hadoop.hbase.classification.InterfaceAudience; 059import org.apache.hadoop.hbase.client.Connection; 060import org.apache.hadoop.hbase.client.ConnectionFactory; @@ -122,670 +122,703 @@ 114 115 public static abstract class Command extends Configured { 116CommandLine cmdline; -117 +117Connection conn; 118Command(Configuration conf) { -119 super(conf); -120} -121 -122public void execute() throws IOException { -123 if (cmdline.hasOption("h") || cmdline.hasOption("help")) { -124printUsage(); -125throw new IOException(INCORRECT_USAGE); -126 } -127} -128 -129protected abstract void printUsage(); -130 } -131 -132 private BackupCommands() { -133throw new AssertionError("Instantiating utility class..."); -134 } -135 -136 public static Command createCommand(Configuration conf, BackupCommand type, CommandLine cmdline) { -137Command cmd = null; -138switch (type) { -139case CREATE: -140 cmd = new CreateCommand(conf, cmdline); -141 break; -142case DESCRIBE: -143 cmd = new DescribeCommand(conf, cmdline); -144 break; -145case PROGRESS: -146 cmd = new ProgressCommand(conf, cmdline); -147 break; -148case DELETE: -149 cmd = new DeleteCommand(conf, cmdline); -150 break; -151case CANCEL: -152 cmd = new CancelCommand(conf, cmdline); -153 break; -154case HISTORY: -155 cmd = new HistoryCommand(conf, cmdline); -156 break; -157case SET: -158 cmd = new BackupSetCommand(conf, cmdline); -159 break; -160case HELP: -161default: -162 cmd = new HelpCommand(conf, cmdline); -163 break; -164} -165return cmd; -166 } -167 -168 static int numOfArgs(String[] args) { -169if (args == null) return 0; -170return args.length; -171 } -172 -173 public static class CreateCommand extends Command { -174 -175CreateCommand(Configuration conf, CommandLine cmdline) { -176 super(conf); -177 this.cmdline = cmdline; -178} -179 -180@Override -181public void execute() throws IOException { -182 super.execute(); -183 if (cmdline == null || cmdline.getArgs() == null) { -184printUsage(); -185throw new IOException(INCORRECT_USAGE); -186 } -187 String[] args = cmdline.getArgs(); -188 if (args.length !=3) { -189printUsage(); -190throw new IOException(INCORRECT_USAGE); -191 } -192 -193 if (!BackupType.FULL.toString().equalsIgnoreCase(args[1]) -194 !BackupType.INCREMENTAL.toString().equalsIgnoreCase(args[1])) { -195System.out.println("ERROR: invalid backup type: " + args[1]); -196printUsage(); -197throw new IOException(INCORRECT_USAGE); -198 } -199 if (!verifyPath(args[2])) { -200System.out.println("ERROR: invalid backup destination: " + args[2]); -201printUsage(); -202throw new IOException(INCORRECT_USAGE); -203 } -204 -205 String tables = null; -206 Configuration conf = getConf() != null ? getConf() : HBaseConfiguration.create(); -207 -208 // Check if we have both: backup set and list of tables -209 if (cmdline.hasOption(OPTION_TABLE) cmdline.hasOption(OPTION_SET)) { -210System.out.println("ERROR: You can specify either backup set or list" -211+ " of tables, but not both"); -212printUsage(); -213throw new IOException(INCORRECT_USAGE); -214 } -215 -216 // Check backup set -217 String setName = null; -218 if (cmdline.hasOption(OPTION_SET)) { -219setName = cmdline.getOptionValue(OPTION_SET); -220tables = getTablesForSet(setName, conf); -221 -222if (tables == null) { -223 System.out.println("ERROR: Backup set '" + setName -224
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/709b8fcc/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html index d77512a..de81692 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html @@ -1543,47 +1543,47 @@ 1535 1536 /** 1537 * Create a new table by cloning the snapshot content. -1538 * @param snapshotName name of the snapshot to be cloned -1539 * @param tableName name of the table where the snapshot will be restored -1540 * @param restoreAcl true to clone acl into newly created table +1538 * +1539 * @param snapshotName name of the snapshot to be cloned +1540 * @param tableName name of the table where the snapshot will be restored 1541 * @throws IOException if a remote or network exception occurs 1542 * @throws TableExistsException if table to be created already exists 1543 * @throws RestoreSnapshotException if snapshot failed to be cloned 1544 * @throws IllegalArgumentException if the specified table has not a valid name 1545 */ -1546 void cloneSnapshot(final String snapshotName, final TableName tableName, final boolean restoreAcl) +1546 void cloneSnapshot(final String snapshotName, final TableName tableName) 1547 throws IOException, TableExistsException, RestoreSnapshotException; 1548 1549 /** -1550 * Create a new table by cloning the snapshot content. -1551 * -1552 * @param snapshotName name of the snapshot to be cloned -1553 * @param tableName name of the table where the snapshot will be restored -1554 * @throws IOException if a remote or network exception occurs -1555 * @throws TableExistsException if table to be created already exists -1556 * @throws RestoreSnapshotException if snapshot failed to be cloned -1557 * @throws IllegalArgumentException if the specified table has not a valid name -1558 */ -1559 void cloneSnapshot(final String snapshotName, final TableName tableName) -1560 throws IOException, TableExistsException, RestoreSnapshotException; -1561 -1562 /** -1563 * Create a new table by cloning the snapshot content, but does not block -1564 * and wait for it be completely cloned. -1565 * You can use Future.get(long, TimeUnit) to wait on the operation to complete. -1566 * It may throw ExecutionException if there was an error while executing the operation -1567 * or TimeoutException in case the wait timeout was not long enough to allow the -1568 * operation to complete. +1550 * Create a new table by cloning the snapshot content, but does not block +1551 * and wait for it be completely cloned. +1552 * You can use Future.get(long, TimeUnit) to wait on the operation to complete. +1553 * It may throw ExecutionException if there was an error while executing the operation +1554 * or TimeoutException in case the wait timeout was not long enough to allow the +1555 * operation to complete. +1556 * +1557 * @param snapshotName name of the snapshot to be cloned +1558 * @param tableName name of the table where the snapshot will be restored +1559 * @throws IOException if a remote or network exception occurs +1560 * @throws TableExistsException if table to be cloned already exists +1561 * @return the result of the async clone snapshot. You can use Future.get(long, TimeUnit) +1562 *to wait on the operation to complete. +1563 */ +1564 FutureVoid cloneSnapshotAsync(final String snapshotName, final TableName tableName) +1565 throws IOException, TableExistsException; +1566 +1567 /** +1568 * Execute a distributed procedure on a cluster. 1569 * -1570 * @param snapshotName name of the snapshot to be cloned -1571 * @param tableName name of the table where the snapshot will be restored -1572 * @throws IOException if a remote or network exception occurs -1573 * @throws TableExistsException if table to be cloned already exists -1574 * @return the result of the async clone snapshot. You can use Future.get(long, TimeUnit) -1575 *to wait on the operation to complete. +1570 * @param signature A distributed procedure is uniquely identified by its signature (default the +1571 * root ZK node name of the procedure). +1572 * @param instance The instance name of the procedure. For some procedures, this parameter is +1573 * optional. +1574 * @param props Property/Value pairs of properties passing to the procedure +1575 * @throws IOException 1576 */ -1577 FutureVoid cloneSnapshotAsync(final String snapshotName, final TableName tableName) -1578 throws IOException, TableExistsException; +1577 void execProcedure(String signature, String instance, MapString, String props) +1578 throws IOException; 1579
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dd7176bf/devapidocs/org/apache/hadoop/hbase/security/access/TablePermission.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/security/access/TablePermission.html b/devapidocs/org/apache/hadoop/hbase/security/access/TablePermission.html index f16a58b..0f910b7 100644 --- a/devapidocs/org/apache/hadoop/hbase/security/access/TablePermission.html +++ b/devapidocs/org/apache/hadoop/hbase/security/access/TablePermission.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -374,14 +374,18 @@ extends readFields(http://docs.oracle.com/javase/8/docs/api/java/io/DataInput.html?is-external=true; title="class or interface in java.io">DataInputin) +void +setTableName(TableNametable) + + boolean tableFieldsEqual(TablePermissionother) - + http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String toString() - + void write(http://docs.oracle.com/javase/8/docs/api/java/io/DataOutput.html?is-external=true; title="class or interface in java.io">DataOutputout) @@ -631,13 +635,22 @@ extends TableNamegetTableName() + + + + + +setTableName +publicvoidsetTableName(TableNametable) + + hasFamily -publicbooleanhasFamily() +publicbooleanhasFamily() @@ -646,7 +659,7 @@ extends getFamily -publicbyte[]getFamily() +publicbyte[]getFamily() @@ -655,7 +668,7 @@ extends hasQualifier -publicbooleanhasQualifier() +publicbooleanhasQualifier() @@ -664,7 +677,7 @@ extends getQualifier -publicbyte[]getQualifier() +publicbyte[]getQualifier() @@ -673,7 +686,7 @@ extends hasNamespace -publicbooleanhasNamespace() +publicbooleanhasNamespace() @@ -682,7 +695,7 @@ extends getNamespace -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetNamespace() +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetNamespace() @@ -691,7 +704,7 @@ extends implies -publicbooleanimplies(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringnamespace, +publicbooleanimplies(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringnamespace, Permission.Actionaction) Checks that a given table operation is authorized by this permission instance. @@ -711,7 +724,7 @@ extends implies -publicbooleanimplies(TableNametable, +publicbooleanimplies(TableNametable, byte[]family, byte[]qualifier, Permission.Actionaction) @@ -737,7 +750,7 @@ extends implies -publicbooleanimplies(TableNametable, +publicbooleanimplies(TableNametable, KeyValuekv, Permission.Actionaction) Checks if this permission grants access to perform the given action on @@ -759,7 +772,7 @@ extends matchesFamily -publicbooleanmatchesFamily(TableNametable, +publicbooleanmatchesFamily(TableNametable, byte[]family, Permission.Actionaction) Returns true if this permission matches the given column @@ -776,7 +789,7 @@ extends matchesFamilyQualifier -publicbooleanmatchesFamilyQualifier(TableNametable, +publicbooleanmatchesFamilyQualifier(TableNametable, byte[]family, byte[]qualifier, Permission.Actionaction) @@ -799,7 +812,7 @@ extends tableFieldsEqual -publicbooleantableFieldsEqual(TablePermissionother) +publicbooleantableFieldsEqual(TablePermissionother) @@ -808,7 +821,7 @@ extends equals -publicbooleanequals(http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Objectobj) +publicbooleanequals(http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Objectobj) Overrides: equalsin classPermission @@ -821,7 +834,7 @@ extends hashCode -publicinthashCode() +publicinthashCode() Overrides: hashCodein classPermission @@
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html index 27e0dee..109b5f3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html @@ -67,157 +67,157 @@ 059import org.apache.hadoop.hbase.CellScanner; 060import org.apache.hadoop.hbase.HConstants; 061import org.apache.hadoop.hbase.Server; -062import org.apache.hadoop.hbase.classification.InterfaceStability; -063import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; -064import org.apache.hadoop.hbase.nio.ByteBuff; -065import org.apache.hadoop.hbase.nio.SingleByteBuff; -066import org.apache.hadoop.hbase.security.AccessDeniedException; -067import org.apache.hadoop.hbase.security.AuthMethod; -068import org.apache.hadoop.hbase.security.HBasePolicyProvider; -069import org.apache.hadoop.hbase.security.SaslStatus; -070import org.apache.hadoop.hbase.security.SaslUtil; -071import org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService; -072import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor; -073import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -074import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -075import org.apache.hadoop.hbase.util.Bytes; -076import org.apache.hadoop.hbase.util.JVM; -077import org.apache.hadoop.hbase.util.Pair; -078import org.apache.hadoop.io.IntWritable; -079import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; -080import org.apache.htrace.TraceInfo; -081 -082/** -083 * An RPC server with Netty4 implementation. -084 * -085 */ -086public class NettyRpcServer extends RpcServer { -087 -088 public static final Log LOG = LogFactory.getLog(NettyRpcServer.class); -089 -090 protected final InetSocketAddress bindAddress; -091 -092 private final CountDownLatch closed = new CountDownLatch(1); -093 private final Channel serverChannel; -094 private final ChannelGroup allChannels = new DefaultChannelGroup(GlobalEventExecutor.INSTANCE);; -095 -096 public NettyRpcServer(final Server server, final String name, -097 final ListBlockingServiceAndInterface services, -098 final InetSocketAddress bindAddress, Configuration conf, -099 RpcScheduler scheduler) throws IOException { -100super(server, name, services, bindAddress, conf, scheduler); -101this.bindAddress = bindAddress; -102boolean useEpoll = useEpoll(conf); -103int workerCount = conf.getInt("hbase.netty.rpc.server.worker.count", -104 Runtime.getRuntime().availableProcessors() / 4); -105EventLoopGroup bossGroup = null; -106EventLoopGroup workerGroup = null; -107if (useEpoll) { -108 bossGroup = new EpollEventLoopGroup(1); -109 workerGroup = new EpollEventLoopGroup(workerCount); -110} else { -111 bossGroup = new NioEventLoopGroup(1); -112 workerGroup = new NioEventLoopGroup(workerCount); -113} -114ServerBootstrap bootstrap = new ServerBootstrap(); -115bootstrap.group(bossGroup, workerGroup); -116if (useEpoll) { -117 bootstrap.channel(EpollServerSocketChannel.class); -118} else { -119 bootstrap.channel(NioServerSocketChannel.class); -120} -121 bootstrap.childOption(ChannelOption.TCP_NODELAY, tcpNoDelay); -122 bootstrap.childOption(ChannelOption.SO_KEEPALIVE, tcpKeepAlive); -123 bootstrap.childOption(ChannelOption.ALLOCATOR, -124 PooledByteBufAllocator.DEFAULT); -125bootstrap.childHandler(new Initializer(maxRequestSize)); -126 -127try { -128 serverChannel = bootstrap.bind(this.bindAddress).sync().channel(); -129 LOG.info("NettyRpcServer bind to address=" + serverChannel.localAddress() -130 + ", hbase.netty.rpc.server.worker.count=" + workerCount -131 + ", useEpoll=" + useEpoll); -132 allChannels.add(serverChannel); -133} catch (InterruptedException e) { -134 throw new InterruptedIOException(e.getMessage()); -135} -136initReconfigurable(conf); -137this.scheduler.init(new RpcSchedulerContext(this)); -138 } -139 -140 private static boolean useEpoll(Configuration conf) { -141// Config to enable native transport. -142boolean epollEnabled = conf.getBoolean("hbase.rpc.server.nativetransport", -143true); -144// Use the faster native epoll transport mechanism on linux if enabled -145return epollEnabled JVM.isLinux() JVM.isAmd64(); -146 } -147 -148 @Override -149 public synchronized void start() { -150if
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/31df4674/devapidocs/org/apache/hadoop/hbase/regionserver/StoreScanner.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/StoreScanner.html b/devapidocs/org/apache/hadoop/hbase/regionserver/StoreScanner.html index 581b976..498fef5 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/StoreScanner.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/StoreScanner.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":9,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":9,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -128,7 +128,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public class StoreScanner +public class StoreScanner extends NonReversedNonLazyKeyValueScanner implements KeyValueScanner, InternalScanner, ChangedReadersObserver Scanner scans both the memstore and the Store. Coalesce KeyValue stream @@ -528,53 +528,54 @@ implements -protected ScanQueryMatcher.MatchCode -optimize(ScanQueryMatcher.MatchCodeqcode, -Cellcell) -See if we should actually SEEK or rather just SKIP to the next Cell (see HBASE-13109). - - - private void parallelSeek(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends KeyValueScannerscanners, Cellkv) Seek storefiles in parallel to optimize IO latency as much as possible - + Cell peek() Look at the next Cell in this scanner, but do not iterate scanner. - + boolean reseek(Cellkv) Reseek the scanner at or after the specified KeyValue. - + protected void resetKVHeap(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends KeyValueScannerscanners, CellComparatorcomparator) - + protected void resetScannerStack(CelllastTopKey) - + boolean seek(Cellkey) Seek the scanner at or after the specified KeyValue. - + protected boolean seekAsDirection(Cellkv) Do a reseek in a normal StoreScanner(scan forward) + +private void +seekOrSkipToNextColumn(Cellcell) + +private void +seekOrSkipToNextRow(Cellcell) + + protected void seekScanners(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends KeyValueScannerscanners, CellseekKey, @@ -583,24 +584,36 @@ implements Seek the specified scanners with the given key - + protected boolean seekToNextRow(Cellc) - + protected http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListKeyValueScanner selectScannersFrom(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">List? extends KeyValueScannerallScanners) Filters the given list of scanners using Bloom filter, time range, and TTL. - + void shipped() Called after a batch of rows scanned and set to be returned to client. - + +protected boolean +trySkipToNextColumn(Cellcell) +See trySkipToNextRow(Cell) + + + +protected boolean +trySkipToNextRow(Cellcell) +See if we should actually SEEK or rather just SKIP to the next Cell (see HBASE-13109). + + + void updateReaders(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListStoreFilesfs) Notify observers. @@ -655,7 +668,7 @@ implements LOG -private static finalorg.apache.commons.logging.Log LOG +private static finalorg.apache.commons.logging.Log LOG @@ -664,7 +677,7 @@ implements store -protected finalStore store +protected finalStore store @@ -673,7 +686,7 @@ implements matcher -protectedScanQueryMatcher matcher +protectedScanQueryMatcher matcher @@ -682,7 +695,7 @@ implements heap -protectedKeyValueHeap heap +protectedKeyValueHeap heap @@ -691,7 +704,7 @@ implements cacheBlocks -protectedboolean cacheBlocks +protectedboolean cacheBlocks @@ -700,7 +713,7 @@ implements countPerRow -protectedlong countPerRow +protectedlong countPerRow @@ -709,7 +722,7 @@ implements storeLimit -protectedint storeLimit +protectedint storeLimit @@ -718,7
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6f2e75f2/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.Converter.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.Converter.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.Converter.html index 6c52543..f3f7a46 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.Converter.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.Converter.html @@ -31,1797 +31,2040 @@ 023import java.util.ArrayList; 024import java.util.Arrays; 025import java.util.Collection; -026import java.util.HashMap; -027import java.util.LinkedList; -028import java.util.List; -029import java.util.Map; -030import java.util.Optional; -031import java.util.concurrent.CompletableFuture; -032import java.util.concurrent.TimeUnit; -033import java.util.concurrent.atomic.AtomicReference; -034import java.util.function.BiConsumer; -035import java.util.regex.Pattern; -036import java.util.stream.Collectors; -037 -038import com.google.common.annotations.VisibleForTesting; -039 -040import io.netty.util.Timeout; -041import io.netty.util.TimerTask; -042import org.apache.commons.logging.Log; -043import org.apache.commons.logging.LogFactory; -044import org.apache.hadoop.hbase.HColumnDescriptor; -045import org.apache.hadoop.hbase.HRegionInfo; -046import org.apache.hadoop.hbase.HRegionLocation; -047import org.apache.hadoop.hbase.MetaTableAccessor; -048import org.apache.hadoop.hbase.MetaTableAccessor.QueryType; -049import org.apache.hadoop.hbase.NotServingRegionException; -050import org.apache.hadoop.hbase.RegionLocations; -051import org.apache.hadoop.hbase.ServerName; -052import org.apache.hadoop.hbase.NamespaceDescriptor; -053import org.apache.hadoop.hbase.HConstants; -054import org.apache.hadoop.hbase.TableExistsException; -055import org.apache.hadoop.hbase.TableName; -056import org.apache.hadoop.hbase.AsyncMetaTableAccessor; -057import org.apache.hadoop.hbase.TableNotFoundException; -058import org.apache.hadoop.hbase.UnknownRegionException; -059import org.apache.hadoop.hbase.classification.InterfaceAudience; -060import org.apache.hadoop.hbase.classification.InterfaceStability; -061import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder; -062import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder; -063import org.apache.hadoop.hbase.client.Scan.ReadType; -064import org.apache.hadoop.hbase.client.replication.ReplicationSerDeHelper; -065import org.apache.hadoop.hbase.client.replication.TableCFs; -066import org.apache.hadoop.hbase.exceptions.DeserializationException; -067import org.apache.hadoop.hbase.ipc.HBaseRpcController; -068import org.apache.hadoop.hbase.quotas.QuotaFilter; -069import org.apache.hadoop.hbase.quotas.QuotaSettings; -070import org.apache.hadoop.hbase.quotas.QuotaTableUtil; -071import org.apache.hadoop.hbase.replication.ReplicationException; -072import org.apache.hadoop.hbase.replication.ReplicationPeerConfig; -073import org.apache.hadoop.hbase.replication.ReplicationPeerDescription; -074import org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback; -075import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -076import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter; -077import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService; -078import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest; -079import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse; -080import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionRequest; -081import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionResponse; -082import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos; -083import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema; -084import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest; -085import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse; -086import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest; -087import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse; -088import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest; -089import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse; -090import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest; -091import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse; -092import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest; -093import
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/efd0601e/checkstyle.rss -- diff --git a/checkstyle.rss b/checkstyle.rss index 3491926..1d3659b 100644 --- a/checkstyle.rss +++ b/checkstyle.rss @@ -25,8 +25,8 @@ under the License. en-us 2007 - 2017 The Apache Software Foundation - File: 2146, - Errors: 14236, + File: 2155, + Errors: 14362, Warnings: 0, Infos: 0 @@ -858,7 +858,7 @@ under the License. - http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.replication.ReplicationLoadSource.java;>org/apache/hadoop/hbase/replication/ReplicationLoadSource.java + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.metrics.MetricRegistriesLoader.java;>org/apache/hadoop/hbase/metrics/MetricRegistriesLoader.java 0 @@ -867,12 +867,12 @@ under the License. 0 - 0 + 2 - http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.metrics.MetricRegistriesLoader.java;>org/apache/hadoop/hbase/metrics/MetricRegistriesLoader.java + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.replication.ReplicationLoadSource.java;>org/apache/hadoop/hbase/replication/ReplicationLoadSource.java 0 @@ -881,7 +881,7 @@ under the License. 0 - 2 + 0 @@ -1180,6 +1180,20 @@ under the License. + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.zookeeper.MetricsZooKeeper.java;>org/apache/hadoop/hbase/zookeeper/MetricsZooKeeper.java + + + 0 + + + 0 + + + 3 + + + + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.util.Triple.java;>org/apache/hadoop/hbase/util/Triple.java @@ -1390,7 +1404,7 @@ under the License. - http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.ExportSnapshotException.java;>org/apache/hadoop/hbase/snapshot/ExportSnapshotException.java + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.replication.regionserver.MetricsReplicationSourceImpl.java;>org/apache/hadoop/hbase/replication/regionserver/MetricsReplicationSourceImpl.java 0 @@ -1404,7 +1418,7 @@ under the License. - http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.replication.regionserver.MetricsReplicationSourceImpl.java;>org/apache/hadoop/hbase/replication/regionserver/MetricsReplicationSourceImpl.java + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.ExportSnapshotException.java;>org/apache/hadoop/hbase/snapshot/ExportSnapshotException.java 0 @@ -2071,7 +2085,7 @@ under the License. 0 - 1 + 2 @@ -3420,7 +3434,7 @@ under the License. - http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.client.AsyncConnectionConfiguration.java;>org/apache/hadoop/hbase/client/AsyncConnectionConfiguration.java + http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.wal.MetricsEditsReplaySource.java;>org/apache/hadoop/hbase/regionserver/wal/MetricsEditsReplaySource.java 0 @@ -3429,12 +3443,12 @@ under the License. 0 - 1 + 0 - http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.wal.MetricsEditsReplaySource.java;>org/apache/hadoop/hbase/regionserver/wal/MetricsEditsReplaySource.java +
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/10601a30/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.AdminRpcCall.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.AdminRpcCall.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.AdminRpcCall.html index be839b7..72853dd 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.AdminRpcCall.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.AdminRpcCall.html @@ -45,1639 +45,1784 @@ 037 038import com.google.common.annotations.VisibleForTesting; 039 -040import org.apache.commons.logging.Log; -041import org.apache.commons.logging.LogFactory; -042import org.apache.hadoop.hbase.HColumnDescriptor; -043import org.apache.hadoop.hbase.HRegionInfo; -044import org.apache.hadoop.hbase.HRegionLocation; -045import org.apache.hadoop.hbase.HTableDescriptor; -046import org.apache.hadoop.hbase.MetaTableAccessor; -047import org.apache.hadoop.hbase.MetaTableAccessor.QueryType; -048import org.apache.hadoop.hbase.NotServingRegionException; -049import org.apache.hadoop.hbase.RegionLocations; -050import org.apache.hadoop.hbase.ServerName; -051import org.apache.hadoop.hbase.NamespaceDescriptor; -052import org.apache.hadoop.hbase.HConstants; -053import org.apache.hadoop.hbase.TableName; -054import org.apache.hadoop.hbase.AsyncMetaTableAccessor; -055import org.apache.hadoop.hbase.TableNotFoundException; -056import org.apache.hadoop.hbase.UnknownRegionException; -057import org.apache.hadoop.hbase.classification.InterfaceAudience; -058import org.apache.hadoop.hbase.classification.InterfaceStability; -059import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder; -060import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder; -061import org.apache.hadoop.hbase.client.Scan.ReadType; -062import org.apache.hadoop.hbase.client.replication.ReplicationSerDeHelper; -063import org.apache.hadoop.hbase.client.replication.TableCFs; -064import org.apache.hadoop.hbase.exceptions.DeserializationException; -065import org.apache.hadoop.hbase.ipc.HBaseRpcController; -066import org.apache.hadoop.hbase.quotas.QuotaFilter; -067import org.apache.hadoop.hbase.quotas.QuotaSettings; -068import org.apache.hadoop.hbase.quotas.QuotaTableUtil; -069import org.apache.hadoop.hbase.replication.ReplicationException; -070import org.apache.hadoop.hbase.replication.ReplicationPeerConfig; -071import org.apache.hadoop.hbase.replication.ReplicationPeerDescription; -072import org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback; -073import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -074import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter; -075import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService; -076import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest; -077import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse; -078import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionRequest; -079import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionResponse; -080import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema; -081import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest; -082import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse; -083import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest; -084import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse; -085import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest; -086import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse; -087import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest; -088import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse; -089import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest; -090import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse; -091import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest; -092import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse; -093import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest; -094import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse; -095import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest; -096import
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/662ea7dc/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.CreateTableProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.CreateTableProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.CreateTableProcedureBiConsumer.html index ac4a9b3..be839b7 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.CreateTableProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.CreateTableProcedureBiConsumer.html @@ -30,212 +30,212 @@ 022import java.io.IOException; 023import java.util.ArrayList; 024import java.util.Arrays; -025import java.util.LinkedList; -026import java.util.List; -027import java.util.Optional; -028import java.util.concurrent.CompletableFuture; -029import java.util.concurrent.TimeUnit; -030import java.util.concurrent.atomic.AtomicReference; -031import java.util.function.BiConsumer; -032import java.util.regex.Pattern; -033 -034import com.google.common.annotations.VisibleForTesting; -035import org.apache.commons.logging.Log; -036import org.apache.commons.logging.LogFactory; -037import org.apache.hadoop.hbase.HColumnDescriptor; -038import org.apache.hadoop.hbase.HRegionInfo; -039import org.apache.hadoop.hbase.HRegionLocation; -040import org.apache.hadoop.hbase.HTableDescriptor; -041import org.apache.hadoop.hbase.MetaTableAccessor; -042import org.apache.hadoop.hbase.MetaTableAccessor.QueryType; -043import org.apache.hadoop.hbase.NotServingRegionException; -044import org.apache.hadoop.hbase.RegionLocations; -045import org.apache.hadoop.hbase.ServerName; -046import org.apache.hadoop.hbase.NamespaceDescriptor; -047import org.apache.hadoop.hbase.HConstants; -048import org.apache.hadoop.hbase.TableName; -049import org.apache.hadoop.hbase.AsyncMetaTableAccessor; -050import org.apache.hadoop.hbase.TableNotFoundException; -051import org.apache.hadoop.hbase.UnknownRegionException; -052import org.apache.hadoop.hbase.classification.InterfaceAudience; -053import org.apache.hadoop.hbase.classification.InterfaceStability; -054import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder; -055import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder; -056import org.apache.hadoop.hbase.client.Scan.ReadType; -057import org.apache.hadoop.hbase.exceptions.DeserializationException; -058import org.apache.hadoop.hbase.ipc.HBaseRpcController; -059import org.apache.hadoop.hbase.quotas.QuotaFilter; -060import org.apache.hadoop.hbase.quotas.QuotaSettings; -061import org.apache.hadoop.hbase.quotas.QuotaTableUtil; -062import org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback; -063import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -064import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter; -065import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService; -066import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest; -067import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse; -068import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionRequest; -069import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionResponse; -070import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema; -071import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest; -072import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse; -073import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest; -074import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse; -075import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest; -076import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse; -077import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest; -078import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse; -079import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest; -080import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse; -081import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest; -082import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse; -083import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest; -084import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse; -085import
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6b4bae59/devapidocs/org/apache/hadoop/hbase/master/HMaster.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html index b9941b0..cea7361 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html +++ b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html @@ -1308,7 +1308,7 @@ implements HRegionServer -abort, addToMovedRegions, addToOnlineRegions, checkFileSystem, cleanMovedRegions, closeAllRegions, closeAndOfflineRegion ForSplitOrMerge, closeRegion, constructRegionServer, convertThrowableToIOE, createClusterConnection, createRegionLoad, createRegionServerS tatusStub, createRegionServerStatusStub, execRegionServerService, getCacheConfig, getChoreService, getClusterConnection, getClusterId, getCompactionPressure, getCompactionRequester, getCompactSplitThread, getConfiguration, getConfigurationManager, getConnection, getCoordinatedStateManager, getExecutorService, getFavoredNodesForRegion, getFileSystem, getFlushPressure, getFlushRequester, getFlushThroughputController, getFromOnlineRegions, getHeapMemoryManager, getInfoServer, getLastSequenceId, getLeases, getMasterAddressTracker, getMetaTableLocator, getMetrics, getMostLoadedRegions, getNonceManager, getNumberOfOnlineRegions, getOnlineRegion, getOnlineRegions, getOnlineRegions, getOnlineRegionsLocalContext, getOnlineTables, getRecoveringRegions, getRegion, getRegionBlockLocations, getRegionByEncodedName, getRegionByEncodedName, getRegionServerAccounting, getRegionServerCoprocessorHost, getRegionServerCoprocessors, getRegionServerMetrics, getRegionServerQuotaManager, getRegionsInTransitionInRS, getReplicationSourceService, getRootDir, getRpcServer, getRSRpcServices, getSecureBulkLoadManager, getStartcode, getThreadWakeFrequency, getWAL, getWALFileSystem, getWalRoller, getWALRootDir, getWALs, handleReportForDutyResponse, isAborted, isOnline, isProcedureFinished, isStopped, isStopping, kill, movedRegionCleanerPeriod, onConfigurationChange, postOpenDeployTasks, postOpenDeployTasks, regionLock, removeFromOnlineRegions, reportRegionStateTransition, reportRegionStateTransition, reportRegionStateTransition, requestRegionSplit, run, setInitLatch, setupClusterConnection, shouldUseThisHostnameInstead, stop, toString, tryRegionServerReport, unassign, updateConfiguration, updateRegionFavoredNodesMapping, waitForServerOnline, walRollRequestFinished +abort, addToMovedRegions, addToOnlineRegions, checkFileSystem, cleanMovedRegions, closeAllRegions, closeAndOfflineRegion ForSplitOrMerge, closeRegion, constructRegionServer, convertThrowableToIOE, createClusterConnection, createRegionLoad, createRegionServerS tatusStub, createRegionServerStatusStub, execRegionServerService, getCacheConfig, getChoreService, getClusterConnection, getClusterId, getCompactionPressure, getCompactionRequester, getCompactSplitThread, getConfiguration, getConfigurationManager, getConnection, getCoordinatedStateManager, getExecutorService, getFavoredNodesForRegion, getFileSystem, getFlushPressure, getFlushRequester, getFlushThroughputController, getFromOnlineRegions, getHeapMemoryManager, getInfoServer, getLastSequenceId, getLeases, getMasterAddressTracker, getMetaTableLocator, getMetrics, getMostLoadedRegions, getNonceManager, getNumberOfOnlineRegions, getOnlineRegion, getOnlineRegions, getOnlineRegions, getOnlineRegionsLocalContext, getOnlineTables, getRecoveringRegions, getRegion, getRegionBlockLocations, getRegionByEncodedName, getRegionByEncodedName, getRegionServerAccounting, getRegionServerCoprocessorHost, getRegionServerCoprocessors, getRegionServerMetrics, getRegionServerQuotaManager, getRegionsInTransitionInRS, getReplicationSourceService, getRootDir, getRpcServer, getRSRpcServices, getSecureBulkLoadManager, getStartcode, getThreadWakeFrequency, getWAL, getWALFileSystem, getWalRoller, getWALRootDir, getWALs, handleReportForDutyResponse, initializeMemStoreChunkCreator, isAborted, isOnline, isProcedureFinished, isStopped, isStopping, kill, movedRegionCleanerPeriod, onConfigurationChange, postOpenDeployTasks, postOpenDeployTasks, regionLock, removeFromOnlineRegions, reportRegionStateTransition, reportRegionStateTransition, reportRegionStateTransition, requestRegionSplit, run, setInitLatch, setupClusterConnection, shouldUseThisHostnameInstead, stop, toString, tryRegionServerReport, unassign, updateConfigu ration, updateRegionFavoredNodesMapping, waitForServerOnline, walRollRequestFinished @@ -2185,7 +2185,7 @@ implements initMobCleaner -privatevoidinitMobCleaner() +privatevoidinitMobCleaner() @@ -2194,7 +2194,7 @@ implements createMetaBootstrap
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2fcc2ae0/devapidocs/org/apache/hadoop/hbase/tool/Canary.RegionServerMonitor.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/tool/Canary.RegionServerMonitor.html b/devapidocs/org/apache/hadoop/hbase/tool/Canary.RegionServerMonitor.html index 4b815e3..1a782ff 100644 --- a/devapidocs/org/apache/hadoop/hbase/tool/Canary.RegionServerMonitor.html +++ b/devapidocs/org/apache/hadoop/hbase/tool/Canary.RegionServerMonitor.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -private static class Canary.RegionServerMonitor +private static class Canary.RegionServerMonitor extends Canary.Monitor @@ -254,7 +254,7 @@ extends allRegions -privateboolean allRegions +privateboolean allRegions @@ -271,7 +271,7 @@ extends RegionServerMonitor -publicRegionServerMonitor(Connectionconnection, +publicRegionServerMonitor(Connectionconnection, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String[]monitorTargets, booleanuseRegExp, Canary.ExtendedSinksink, @@ -294,7 +294,7 @@ extends getSink -privateCanary.ExtendedSinkgetSink() +privateCanary.ExtendedSinkgetSink() @@ -303,7 +303,7 @@ extends run -publicvoidrun() +publicvoidrun() Specified by: http://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true#run--; title="class or interface in java.lang">runin interfacehttp://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true; title="class or interface in java.lang">Runnable @@ -318,7 +318,7 @@ extends checkNoTableNames -privatebooleancheckNoTableNames() +privatebooleancheckNoTableNames() @@ -327,7 +327,7 @@ extends monitorRegionServers -privatevoidmonitorRegionServers(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListHRegionInforsAndRMap) +privatevoidmonitorRegionServers(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListHRegionInforsAndRMap) @@ -336,7 +336,7 @@ extends filterRegionServerByName -privatehttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListHRegionInfofilterRegionServerByName() +privatehttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListHRegionInfofilterRegionServerByName() @@ -345,7 +345,7 @@ extends getAllRegionServerByName -privatehttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListHRegionInfogetAllRegionServerByName() +privatehttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListHRegionInfogetAllRegionServerByName() @@ -354,7 +354,7 @@ extends doFilterRegionServerByName -privatehttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e4348f53/devapidocs/org/apache/hadoop/hbase/regionserver/BusyRegionSplitPolicy.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/BusyRegionSplitPolicy.html b/devapidocs/org/apache/hadoop/hbase/regionserver/BusyRegionSplitPolicy.html index 9fa46dc..5bc3c7f 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/BusyRegionSplitPolicy.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/BusyRegionSplitPolicy.html @@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab"; PrevClass -NextClass +NextClass Frames @@ -563,7 +563,7 @@ extends PrevClass -NextClass +NextClass Frames http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e4348f53/devapidocs/org/apache/hadoop/hbase/regionserver/ByteBufferChunkCell.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/ByteBufferChunkCell.html b/devapidocs/org/apache/hadoop/hbase/regionserver/ByteBufferChunkCell.html new file mode 100644 index 000..f08b4b2 --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/ByteBufferChunkCell.html @@ -0,0 +1,353 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +ByteBufferChunkCell (Apache HBase 2.0.0-SNAPSHOT API) + + + + + +var methods = {"i0":10}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Field| +Constr| +Method + + +Detail: +Field| +Constr| +Method + + + + + + + + +org.apache.hadoop.hbase.regionserver +Class ByteBufferChunkCell + + + +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.ByteBufferCell + + +org.apache.hadoop.hbase.ByteBufferKeyValue + + +org.apache.hadoop.hbase.regionserver.ByteBufferChunkCell + + + + + + + + + + + +All Implemented Interfaces: +http://docs.oracle.com/javase/8/docs/api/java/lang/Cloneable.html?is-external=true; title="class or interface in java.lang">Cloneable, Cell, ExtendedCell, HeapSize, SettableSequenceId, SettableTimestamp + + + +@InterfaceAudience.Private +public class ByteBufferChunkCell +extends ByteBufferKeyValue +ByteBuffer based cell which has the chunkid at the 0th offset + +See Also: +MemStoreLAB + + + + + + + + + + + + +Field Summary + + + + +Fields inherited from classorg.apache.hadoop.hbase.ByteBufferKeyValue +buf, FIXED_OVERHEAD, length, offset + + + + + +Fields inherited from interfaceorg.apache.hadoop.hbase.ExtendedCell +CELL_NOT_BASED_ON_CHUNK + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +ByteBufferChunkCell(http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true; title="class or interface in java.nio">ByteBufferbuf, + intoffset, + intlength) + + +ByteBufferChunkCell(http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true; title="class or interface in java.nio">ByteBufferbuf, + intoffset, + intlength, + longseqId) + + + + + + + + + +Method Summary + +All MethodsInstance MethodsConcrete Methods + +Modifier and Type +Method and Description + + +int +getChunkId() +Extracts the id of the backing bytebuffer of this cell if it was obtained from fixed sized + chunks as in case of MemstoreLAB + + + + + + + +Methods inherited from classorg.apache.hadoop.hbase.ByteBufferKeyValue +deepClone, equals, getBuffer, getFamilyArray, getFamilyByteBuffer, getFamilyLength, getFamilyOffset, getFamilyPosition, getOffset, getQualifierArray, getQualifierByteBuffer, getQualifierLength, getQualifierOffset, getQualifierPosition, getRowArray, getRowByteBuffer, getRowLength, getRowOffset, getRowPosition, getSequenceId, getSerializedSize, getTagsArray, getTagsByteBuffer, getTagsLength, getTagsOffset, getTagsPosition, getTimestamp, getTypeByte, getValueArray, getValueByteBuffer, getValueLength, getValueOffset, getValuePosition, hashCode, heapSize,
[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e57d1b63/apidocs/org/apache/hadoop/hbase/TableInfoMissingException.html -- diff --git a/apidocs/org/apache/hadoop/hbase/TableInfoMissingException.html b/apidocs/org/apache/hadoop/hbase/TableInfoMissingException.html deleted file mode 100644 index f13e6ec..000 --- a/apidocs/org/apache/hadoop/hbase/TableInfoMissingException.html +++ /dev/null @@ -1,313 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd;> - - - - - -TableInfoMissingException (Apache HBase 2.0.0-SNAPSHOT API) - - - - - - - -JavaScript is disabled on your browser. - - - - - -Skip navigation links - - - - -Overview -Package -Class -Use -Tree -Deprecated -Index -Help - - - - -PrevClass -NextClass - - -Frames -NoFrames - - -AllClasses - - - - - - - -Summary: -Nested| -Field| -Constr| -Method - - -Detail: -Field| -Constr| -Method - - - - - - - - -org.apache.hadoop.hbase -Class TableInfoMissingException - - - -http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object - - -http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">java.lang.Throwable - - -http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true; title="class or interface in java.lang">java.lang.Exception - - -http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">java.io.IOException - - -org.apache.hadoop.hbase.HBaseIOException - - -org.apache.hadoop.hbase.TableInfoMissingException - - - - - - - - - - - - - - - -All Implemented Interfaces: -http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable - - - -@InterfaceAudience.Public -public class TableInfoMissingException -extends HBaseIOException -Failed to find .tableinfo file under table dir - -See Also: -Serialized Form - - - - - - - - - - - - -Constructor Summary - -Constructors - -Constructor and Description - - -TableInfoMissingException() - - -TableInfoMissingException(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringmessage) - - -TableInfoMissingException(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringmessage, - http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwablet) - - -TableInfoMissingException(http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwablet) - - - - - - - - - -Method Summary - - - - -Methods inherited from classjava.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwable -http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#addSuppressed-java.lang.Throwable-; title="class or interface in java.lang">addSuppressed, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#fillInStackTrace--; title="class or interface in java.lang">fillInStackTrace, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getCause--; title="class or interface in java.lang">getCause, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getLocalizedMessage--; title="class or interface in java.lang">getLocalizedMessage, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getMessage--; title="class or interface in java.lang">getMessage, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getStackTrace--; title="class or inter face in java.lang">getStackTrace, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getSuppressed--; title="class or interface in java.lang">getSuppressed, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#initCause-java.lang.Throwable-; title="class or interface in java.lang">initCause, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#printStackTrace--; title="class or interface in java.lang">printStackTrace,