[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-08-11 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/346adc37/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html
index b3d18d1..1c9f960 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-summary.html
@@ -196,7 +196,7 @@
 
 UnassignProcedure
 
-Procedure that describe the unassignment of a single 
region.
+Procedure that describes the unassignment of a single 
region.
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/346adc37/devapidocs/org/apache/hadoop/hbase/master/assignment/package-use.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-use.html 
b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-use.html
index e94e8e4..12d4817 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/assignment/package-use.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/package-use.html
@@ -205,7 +205,7 @@
 
 
 UnassignProcedure
-Procedure that describe the unassignment of a single 
region.
+Procedure that describes the unassignment of a single 
region.
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/346adc37/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
index 9b1d336..86c040e 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
@@ -312,11 +312,11 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective
-org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus
+org.apache.hadoop.hbase.master.RegionState.State
 org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode
+org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus
+org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective
 org.apache.hadoop.hbase.master.MetricsMasterSourceFactoryImpl.FactoryStorage
-org.apache.hadoop.hbase.master.RegionState.State
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/346adc37/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashException.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashException.html 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashException.html
index 066581d..89087f7 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashException.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/ServerCrashException.html
@@ -134,10 +134,11 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class ServerCrashException
+public class ServerCrashException
 extends HBaseIOException
 Passed as Exception by ServerCrashProcedure
- notifying on-going RIT that server has failed.
+ notifying on-going RIT that server has failed. This exception is less an 
error-condition than
+ it is a signal to waiting procedures that they can now proceed.
 
 See Also:
 Serialized
 Form
@@ -240,7 +241,7 @@ extends 
 
 procId
-private finallong procId
+private finallong procId
 
 
 
@@ -249,7 +250,7 @@ extends 
 
 serverName
-private finalServerName serverName
+private finalServerName serverName
 
 
 
@@ -266,7 +267,7 @@ extends 
 
 ServerCrashException
-publicServerCrashException(longprocId,
+publicServerCrashException(longprocId,
 ServerNameserverName)
 
 Parameters:
@@ -288,7 +289,7 @@ extends 
 
 getMessage
-publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringgetMessage()
+publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringgetMessage()
 
 Overrides:
 http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getMessage--;
 title="class or interface in java.lang">getMessagein 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-08-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2aec596e/devapidocs/org/apache/hadoop/hbase/class-use/CellComparator.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/CellComparator.html 
b/devapidocs/org/apache/hadoop/hbase/class-use/CellComparator.html
index c5d849a..6119d39 100644
--- a/devapidocs/org/apache/hadoop/hbase/class-use/CellComparator.html
+++ b/devapidocs/org/apache/hadoop/hbase/class-use/CellComparator.html
@@ -253,12 +253,12 @@
 
 
 int
-RowIndexSeekerV1.compareKey(CellComparatorcomparator,
+BufferedDataBlockEncoder.BufferedEncodedSeeker.compareKey(CellComparatorcomparator,
   Cellkey)
 
 
 int
-BufferedDataBlockEncoder.BufferedEncodedSeeker.compareKey(CellComparatorcomparator,
+RowIndexSeekerV1.compareKey(CellComparatorcomparator,
   Cellkey)
 
 
@@ -270,27 +270,27 @@
 
 
 DataBlockEncoder.EncodedSeeker
-RowIndexCodecV1.createSeeker(CellComparatorcomparator,
+CopyKeyDataBlockEncoder.createSeeker(CellComparatorcomparator,
 HFileBlockDecodingContextdecodingCtx)
 
 
 DataBlockEncoder.EncodedSeeker
-CopyKeyDataBlockEncoder.createSeeker(CellComparatorcomparator,
+PrefixKeyDeltaEncoder.createSeeker(CellComparatorcomparator,
 HFileBlockDecodingContextdecodingCtx)
 
 
 DataBlockEncoder.EncodedSeeker
-DiffKeyDeltaEncoder.createSeeker(CellComparatorcomparator,
+FastDiffDeltaEncoder.createSeeker(CellComparatorcomparator,
 HFileBlockDecodingContextdecodingCtx)
 
 
 DataBlockEncoder.EncodedSeeker
-FastDiffDeltaEncoder.createSeeker(CellComparatorcomparator,
+DiffKeyDeltaEncoder.createSeeker(CellComparatorcomparator,
 HFileBlockDecodingContextdecodingCtx)
 
 
 DataBlockEncoder.EncodedSeeker
-PrefixKeyDeltaEncoder.createSeeker(CellComparatorcomparator,
+RowIndexCodecV1.createSeeker(CellComparatorcomparator,
 HFileBlockDecodingContextdecodingCtx)
 
 
@@ -328,9 +328,9 @@
 
 
 
-protected CellComparator
-HFileWriterImpl.comparator
-Key comparator.
+private CellComparator
+HFileBlockIndex.CellBasedKeyBlockIndexReader.comparator
+Needed doing lookup on blocks.
 
 
 
@@ -344,9 +344,9 @@
 
 
 
-private CellComparator
-HFileBlockIndex.CellBasedKeyBlockIndexReader.comparator
-Needed doing lookup on blocks.
+protected CellComparator
+HFileWriterImpl.comparator
+Key comparator.
 
 
 
@@ -530,8 +530,8 @@
 StripeStoreFileManager.cellComparator
 
 
-protected CellComparator
-StripeMultiFileWriter.comparator
+private CellComparator
+StoreFileWriter.Builder.comparator
 
 
 protected CellComparator
@@ -539,36 +539,36 @@
 
 
 private CellComparator
-Segment.comparator
+AbstractMemStore.comparator
 
 
 private CellComparator
-ScanInfo.comparator
+Segment.comparator
 
 
-private CellComparator
-StoreFileWriter.Builder.comparator
+protected CellComparator
+HRegion.RegionScannerImpl.comparator
 
 
 protected CellComparator
-HRegion.RegionScannerImpl.comparator
+StripeMultiFileWriter.comparator
 
 
 private CellComparator
-AbstractMemStore.comparator
+ScanInfo.comparator
 
 
 protected CellComparator
 HStore.comparator
 
 
-protected CellComparator
-KeyValueHeap.KVScannerComparator.kvComparator
-
-
 private CellComparator
 DefaultStoreFileManager.kvComparator
 
+
+protected CellComparator
+KeyValueHeap.KVScannerComparator.kvComparator
+
 
 
 
@@ -589,38 +589,38 @@
 HRegion.getCellComparator()
 
 
-CellComparator
-StoreFileReader.getComparator()
+(package private) CellComparator
+StoreFileScanner.getComparator()
 
 
 protected CellComparator
-Segment.getComparator()
-Returns the Cell comparator used by this segment
-
+AbstractMemStore.getComparator()
 
 
 CellComparator
-ScanInfo.getComparator()
+StoreFileReader.getComparator()
 
 
 CellComparator
 KeyValueHeap.KVScannerComparator.getComparator()
 
 
-CellComparator
-Store.getComparator()
+protected CellComparator
+Segment.getComparator()
+Returns the Cell comparator used by this segment
+
 
 
-protected CellComparator
-AbstractMemStore.getComparator()
+CellComparator
+ScanInfo.getComparator()
 
 
 CellComparator
-HStore.getComparator()
+Store.getComparator()
 
 
-(package private) CellComparator
-StoreFileScanner.getComparator()
+CellComparator
+HStore.getComparator()
 
 
 
@@ -663,21 +663,21 @@
 
 
 protected void
-DateTieredStoreEngine.createComponents(org.apache.hadoop.conf.Configurationconf,
+StripeStoreEngine.createComponents(org.apache.hadoop.conf.Configurationconf,
 Storestore,
-CellComparatorkvComparator)
+CellComparatorcomparator)
 
 
 protected void
-DefaultStoreEngine.createComponents(org.apache.hadoop.conf.Configurationconf,
+DateTieredStoreEngine.createComponents(org.apache.hadoop.conf.Configurationconf,
 Storestore,
 CellComparatorkvComparator)
 
 
 protected void
-StripeStoreEngine.createComponents(org.apache.hadoop.conf.Configurationconf,
+DefaultStoreEngine.createComponents(org.apache.hadoop.conf.Configurationconf,
  

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-31 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1837997e/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MergeTableRegionProcedureBiConsumer.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MergeTableRegionProcedureBiConsumer.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MergeTableRegionProcedureBiConsumer.html
index 35d5549..7f42873 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MergeTableRegionProcedureBiConsumer.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MergeTableRegionProcedureBiConsumer.html
@@ -115,2816 +115,2814 @@
 107import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse;
 108import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterRequest;
 109import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.RollWALWriterResponse;
-110import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionRequest;
-111import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionResponse;
-112import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerRequest;
-113import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.StopServerResponse;
-114import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationRequest;
-115import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.UpdateConfigurationResponse;
-116import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription;
-117import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema;
-118import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType;
-119import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest;
-120import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse;
-121import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest;
-122import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse;
-123import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest;
-124import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse;
-125import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest;
-126import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse;
-127import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest;
-128import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse;
-129import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest;
-130import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse;
-131import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest;
-132import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse;
-133import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest;
-134import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse;
-135import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest;
-136import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse;
-137import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest;
-138import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse;
-139import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest;
-140import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse;
-141import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest;
-142import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse;
-143import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest;
-144import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse;
-145import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest;
-146import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse;
-147import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest;
-148import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse;
-149import 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a2b2dd19/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
index 5c95397..860416b 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RowLockContext.html
@@ -293,7944 +293,7962 @@
 285  final AtomicLong compactionsFailed = 
new AtomicLong(0L);
 286  final AtomicLong 
compactionNumFilesCompacted = new AtomicLong(0L);
 287  final AtomicLong 
compactionNumBytesCompacted = new AtomicLong(0L);
-288
-289  private final WAL wal;
-290  private final HRegionFileSystem fs;
-291  protected final Configuration conf;
-292  private final Configuration baseConf;
-293  private final int 
rowLockWaitDuration;
-294  static final int 
DEFAULT_ROWLOCK_WAIT_DURATION = 3;
-295
-296  // The internal wait duration to 
acquire a lock before read/update
-297  // from the region. It is not per row. 
The purpose of this wait time
-298  // is to avoid waiting a long time 
while the region is busy, so that
-299  // we can release the IPC handler soon 
enough to improve the
-300  // availability of the region server. 
It can be adjusted by
-301  // tuning configuration 
"hbase.busy.wait.duration".
-302  final long busyWaitDuration;
-303  static final long 
DEFAULT_BUSY_WAIT_DURATION = HConstants.DEFAULT_HBASE_RPC_TIMEOUT;
-304
-305  // If updating multiple rows in one 
call, wait longer,
-306  // i.e. waiting for busyWaitDuration * 
# of rows. However,
-307  // we can limit the max multiplier.
-308  final int maxBusyWaitMultiplier;
-309
-310  // Max busy wait duration. There is no 
point to wait longer than the RPC
-311  // purge timeout, when a RPC call will 
be terminated by the RPC engine.
-312  final long maxBusyWaitDuration;
-313
-314  // Max cell size. If nonzero, the 
maximum allowed size for any given cell
-315  // in bytes
-316  final long maxCellSize;
-317
-318  // negative number indicates infinite 
timeout
-319  static final long 
DEFAULT_ROW_PROCESSOR_TIMEOUT = 60 * 1000L;
-320  final ExecutorService 
rowProcessorExecutor = Executors.newCachedThreadPool();
-321
-322  private final 
ConcurrentHashMapRegionScanner, Long scannerReadPoints;
+288  final AtomicLong compactionsQueued = 
new AtomicLong(0L);
+289  final AtomicLong flushesQueued = new 
AtomicLong(0L);
+290
+291  private final WAL wal;
+292  private final HRegionFileSystem fs;
+293  protected final Configuration conf;
+294  private final Configuration baseConf;
+295  private final int 
rowLockWaitDuration;
+296  static final int 
DEFAULT_ROWLOCK_WAIT_DURATION = 3;
+297
+298  // The internal wait duration to 
acquire a lock before read/update
+299  // from the region. It is not per row. 
The purpose of this wait time
+300  // is to avoid waiting a long time 
while the region is busy, so that
+301  // we can release the IPC handler soon 
enough to improve the
+302  // availability of the region server. 
It can be adjusted by
+303  // tuning configuration 
"hbase.busy.wait.duration".
+304  final long busyWaitDuration;
+305  static final long 
DEFAULT_BUSY_WAIT_DURATION = HConstants.DEFAULT_HBASE_RPC_TIMEOUT;
+306
+307  // If updating multiple rows in one 
call, wait longer,
+308  // i.e. waiting for busyWaitDuration * 
# of rows. However,
+309  // we can limit the max multiplier.
+310  final int maxBusyWaitMultiplier;
+311
+312  // Max busy wait duration. There is no 
point to wait longer than the RPC
+313  // purge timeout, when a RPC call will 
be terminated by the RPC engine.
+314  final long maxBusyWaitDuration;
+315
+316  // Max cell size. If nonzero, the 
maximum allowed size for any given cell
+317  // in bytes
+318  final long maxCellSize;
+319
+320  // negative number indicates infinite 
timeout
+321  static final long 
DEFAULT_ROW_PROCESSOR_TIMEOUT = 60 * 1000L;
+322  final ExecutorService 
rowProcessorExecutor = Executors.newCachedThreadPool();
 323
-324  /**
-325   * The sequence ID that was 
enLongAddered when this region was opened.
-326   */
-327  private long openSeqNum = 
HConstants.NO_SEQNUM;
-328
-329  /**
-330   * The default setting for whether to 
enable on-demand CF loading for
-331   * scan requests to this region. 
Requests can override it.
-332   */
-333  private boolean 
isLoadingCfsOnDemandDefault = false;
-334
-335  private final AtomicInteger 
majorInProgress = new AtomicInteger(0);
-336  private final AtomicInteger 
minorInProgress = new AtomicInteger(0);
-337
-338  //
-339  // Context: During replay we want to 
ensure that we do not lose any data. So, we
-340  // have to be conservative in how we 
replay wals. For each store, we calculate
-341  // the maxSeqId up to which 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-26 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/21766f4a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/class-use/MinorCompactionScanQueryMatcher.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/class-use/MinorCompactionScanQueryMatcher.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/class-use/MinorCompactionScanQueryMatcher.html
index 4d1e498..d53da57 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/class-use/MinorCompactionScanQueryMatcher.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/class-use/MinorCompactionScanQueryMatcher.html
@@ -72,7 +72,49 @@
 
 Uses of 
Classorg.apache.hadoop.hbase.regionserver.querymatcher.MinorCompactionScanQueryMatcher
 
-No usage of 
org.apache.hadoop.hbase.regionserver.querymatcher.MinorCompactionScanQueryMatcher
+
+
+
+
+Packages that use MinorCompactionScanQueryMatcher
+
+Package
+Description
+
+
+
+org.apache.hadoop.hbase.regionserver.querymatcher
+
+
+
+
+
+
+
+
+
+
+Uses of MinorCompactionScanQueryMatcher
 in org.apache.hadoop.hbase.regionserver.querymatcher
+
+Subclasses of MinorCompactionScanQueryMatcher
 in org.apache.hadoop.hbase.regionserver.querymatcher
+
+Modifier and Type
+Class and Description
+
+
+
+class
+IncludeAllCompactionQueryMatcher
+A compaction query matcher that always return INCLUDE and 
drops nothing.
+
+
+
+
+
+
+
+
+
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/21766f4a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/class-use/NewVersionBehaviorTracker.DeleteVersionsNode.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/class-use/NewVersionBehaviorTracker.DeleteVersionsNode.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/class-use/NewVersionBehaviorTracker.DeleteVersionsNode.html
new file mode 100644
index 000..ae2582c
--- /dev/null
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/class-use/NewVersionBehaviorTracker.DeleteVersionsNode.html
@@ -0,0 +1,204 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+
+Uses of Class 
org.apache.hadoop.hbase.regionserver.querymatcher.NewVersionBehaviorTracker.DeleteVersionsNode
 (Apache HBase 3.0.0-SNAPSHOT API)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev
+Next
+
+
+Frames
+NoFrames
+
+
+AllClasses
+
+
+
+
+
+
+
+
+
+
+Uses of 
Classorg.apache.hadoop.hbase.regionserver.querymatcher.NewVersionBehaviorTracker.DeleteVersionsNode
+
+
+
+
+
+Packages that use NewVersionBehaviorTracker.DeleteVersionsNode
+
+Package
+Description
+
+
+
+org.apache.hadoop.hbase.regionserver.querymatcher
+
+
+
+org.apache.hadoop.hbase.security.visibility
+
+
+
+
+
+
+
+
+
+
+Uses of NewVersionBehaviorTracker.DeleteVersionsNode
 in org.apache.hadoop.hbase.regionserver.querymatcher
+
+Fields in org.apache.hadoop.hbase.regionserver.querymatcher
 with type parameters of type NewVersionBehaviorTracker.DeleteVersionsNode
+
+Modifier and Type
+Field and Description
+
+
+
+protected http://docs.oracle.com/javase/8/docs/api/java/util/NavigableMap.html?is-external=true;
 title="class or interface in java.util">NavigableMaphttp://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
 title="class or interface in java.lang">Long,NewVersionBehaviorTracker.DeleteVersionsNode
+NewVersionBehaviorTracker.delColMap
+
+
+protected http://docs.oracle.com/javase/8/docs/api/java/util/NavigableMap.html?is-external=true;
 title="class or interface in java.util">NavigableMaphttp://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
 title="class or interface in java.lang">Long,NewVersionBehaviorTracker.DeleteVersionsNode
+NewVersionBehaviorTracker.delFamMap
+
+
+
+
+Methods in org.apache.hadoop.hbase.regionserver.querymatcher
 that return NewVersionBehaviorTracker.DeleteVersionsNode
+
+Modifier and Type
+Method and Description
+
+
+
+protected NewVersionBehaviorTracker.DeleteVersionsNode
+NewVersionBehaviorTracker.DeleteVersionsNode.getDeepCopy()
+
+
+
+
+
+
+
+Uses of NewVersionBehaviorTracker.DeleteVersionsNode
 in org.apache.hadoop.hbase.security.visibility
+
+Subclasses of 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-24 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2d5075d7/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
index 22817f5..582d47a 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
@@ -692,19 +692,19 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.regionserver.ScanType
-org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection
-org.apache.hadoop.hbase.regionserver.BloomType
+org.apache.hadoop.hbase.regionserver.RegionOpeningState
+org.apache.hadoop.hbase.regionserver.ScannerContext.NextState
 org.apache.hadoop.hbase.regionserver.Region.FlushResult.Result
+org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status
 org.apache.hadoop.hbase.regionserver.Region.Operation
+org.apache.hadoop.hbase.regionserver.BloomType
+org.apache.hadoop.hbase.regionserver.FlushType
+org.apache.hadoop.hbase.regionserver.ScanType
 org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope
-org.apache.hadoop.hbase.regionserver.ScannerContext.NextState
-org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage
 org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType
 org.apache.hadoop.hbase.regionserver.MemStoreCompactor.Action
-org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status
-org.apache.hadoop.hbase.regionserver.RegionOpeningState
-org.apache.hadoop.hbase.regionserver.FlushType
+org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection
+org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage
 org.apache.hadoop.hbase.regionserver.StoreScanner.StoreScannerCompactionRace
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2d5075d7/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html
index 7cba16c..33cb6b6 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/wal/package-tree.html
@@ -248,8 +248,8 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.regionserver.wal.RingBufferTruck.Type
 org.apache.hadoop.hbase.regionserver.wal.ProtobufLogReader.WALHdrResult
+org.apache.hadoop.hbase.regionserver.wal.RingBufferTruck.Type
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2d5075d7/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html
index 1185ae8..2680226 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html
@@ -214,8 +214,8 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.replication.regionserver.ReplicationSourceShipper.WorkerState
 org.apache.hadoop.hbase.replication.regionserver.MetricsReplicationSourceFactoryImpl.SourceHolder
+org.apache.hadoop.hbase.replication.regionserver.ReplicationSourceShipper.WorkerState
 
 
 


[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0383a9c2/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.html
index 6c200a1..e6f8c2e 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.html
@@ -1350,726 +1350,719 @@
 1342  }
 1343
 1344  @Override
-1345  public 
MasterProtos.DispatchMergingRegionsResponse dispatchMergingRegions(
-1346  RpcController controller, 
MasterProtos.DispatchMergingRegionsRequest request)
-1347  throws ServiceException {
-1348return 
stub.dispatchMergingRegions(controller, request);
-1349  }
-1350
-1351  @Override
-1352  public 
MasterProtos.AssignRegionResponse assignRegion(RpcController controller,
-1353  
MasterProtos.AssignRegionRequest request) throws ServiceException {
-1354return 
stub.assignRegion(controller, request);
-1355  }
-1356
-1357  @Override
-1358  public 
MasterProtos.UnassignRegionResponse unassignRegion(RpcController controller,
-1359  
MasterProtos.UnassignRegionRequest request) throws ServiceException {
-1360return 
stub.unassignRegion(controller, request);
-1361  }
-1362
-1363  @Override
-1364  public 
MasterProtos.OfflineRegionResponse offlineRegion(RpcController controller,
-1365  
MasterProtos.OfflineRegionRequest request) throws ServiceException {
-1366return 
stub.offlineRegion(controller, request);
-1367  }
-1368
-1369  @Override
-1370  public 
MasterProtos.SplitTableRegionResponse splitRegion(RpcController controller,
-1371  
MasterProtos.SplitTableRegionRequest request) throws ServiceException {
-1372return 
stub.splitRegion(controller, request);
-1373  }
-1374
-1375  @Override
-1376  public 
MasterProtos.DeleteTableResponse deleteTable(RpcController controller,
-1377  
MasterProtos.DeleteTableRequest request) throws ServiceException {
-1378return 
stub.deleteTable(controller, request);
-1379  }
-1380
-1381  @Override
-1382  public 
MasterProtos.TruncateTableResponse truncateTable(RpcController controller,
-1383  
MasterProtos.TruncateTableRequest request) throws ServiceException {
-1384return 
stub.truncateTable(controller, request);
-1385  }
-1386
-1387  @Override
-1388  public 
MasterProtos.EnableTableResponse enableTable(RpcController controller,
-1389  
MasterProtos.EnableTableRequest request) throws ServiceException {
-1390return 
stub.enableTable(controller, request);
-1391  }
-1392
-1393  @Override
-1394  public 
MasterProtos.DisableTableResponse disableTable(RpcController controller,
-1395  
MasterProtos.DisableTableRequest request) throws ServiceException {
-1396return 
stub.disableTable(controller, request);
-1397  }
-1398
-1399  @Override
-1400  public 
MasterProtos.ModifyTableResponse modifyTable(RpcController controller,
-1401  
MasterProtos.ModifyTableRequest request) throws ServiceException {
-1402return 
stub.modifyTable(controller, request);
-1403  }
-1404
-1405  @Override
-1406  public 
MasterProtos.CreateTableResponse createTable(RpcController controller,
-1407  
MasterProtos.CreateTableRequest request) throws ServiceException {
-1408return 
stub.createTable(controller, request);
-1409  }
-1410
-1411  @Override
-1412  public 
MasterProtos.ShutdownResponse shutdown(RpcController controller,
-1413  MasterProtos.ShutdownRequest 
request) throws ServiceException {
-1414return stub.shutdown(controller, 
request);
-1415  }
-1416
-1417  @Override
-1418  public 
MasterProtos.StopMasterResponse stopMaster(RpcController controller,
-1419  MasterProtos.StopMasterRequest 
request) throws ServiceException {
-1420return 
stub.stopMaster(controller, request);
+1345  public 
MasterProtos.AssignRegionResponse assignRegion(RpcController controller,
+1346  
MasterProtos.AssignRegionRequest request) throws ServiceException {
+1347return 
stub.assignRegion(controller, request);
+1348  }
+1349
+1350  @Override
+1351  public 
MasterProtos.UnassignRegionResponse unassignRegion(RpcController controller,
+1352  
MasterProtos.UnassignRegionRequest request) throws ServiceException {
+1353return 
stub.unassignRegion(controller, request);
+1354  }
+1355
+1356  @Override
+1357  public 
MasterProtos.OfflineRegionResponse offlineRegion(RpcController controller,
+1358  
MasterProtos.OfflineRegionRequest request) throws 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-22 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f391bcef/devapidocs/index-all.html
--
diff --git a/devapidocs/index-all.html b/devapidocs/index-all.html
index 8b8f33b..ad2b82a 100644
--- a/devapidocs/index-all.html
+++ b/devapidocs/index-all.html
@@ -778,6 +778,8 @@
 
 Action(Row,
 int) - Constructor for class org.apache.hadoop.hbase.client.Action
 
+Action(Row,
 int, int) - Constructor for class org.apache.hadoop.hbase.client.Action
+
 Action(Action,
 int) - Constructor for class org.apache.hadoop.hbase.client.Action
 
 Creates an action for a particular replica from original 
action.
@@ -8929,7 +8931,7 @@
 
 This class is used to unify HTable calls with AsyncProcess 
Framework.
 
-CancellableRegionServerCallable(Connection,
 TableName, byte[], RpcController, int, RetryingTimeTracker) - 
Constructor for class org.apache.hadoop.hbase.client.CancellableRegionServerCallable
+CancellableRegionServerCallable(Connection,
 TableName, byte[], RpcController, int, RetryingTimeTracker, int) - 
Constructor for class org.apache.hadoop.hbase.client.CancellableRegionServerCallable
 
 cancellationCbs
 - Variable in class org.apache.hadoop.hbase.ipc.HBaseRpcControllerImpl
 
@@ -12006,7 +12008,7 @@
 
 A RegionServerCallable set to use the Client protocol.
 
-ClientServiceCallable(Connection,
 TableName, byte[], RpcController) - Constructor for class 
org.apache.hadoop.hbase.client.ClientServiceCallable
+ClientServiceCallable(Connection,
 TableName, byte[], RpcController, int) - Constructor for class 
org.apache.hadoop.hbase.client.ClientServiceCallable
 
 ClientSideRegionScanner - Class in org.apache.hadoop.hbase.client
 
@@ -40151,6 +40153,8 @@
 
 getMaxPoolSize()
 - Method in interface org.apache.hadoop.hbase.thrift.IncrementCoalescerMBean
 
+getMaxPriority()
 - Method in class org.apache.hadoop.hbase.client.RowMutations
+
 getMaxProcId()
 - Method in class org.apache.hadoop.hbase.procedure2.store.wal.ProcedureWALFile
 
 getMaxProcId()
 - Method in class org.apache.hadoop.hbase.procedure2.store.wal.ProcedureWALFormatReader.WalProcedureMap
@@ -43284,6 +43288,14 @@
 
 Get the user name from a principal
 
+getPriority()
 - Method in class org.apache.hadoop.hbase.client.Action
+
+getPriority()
 - Method in class org.apache.hadoop.hbase.client.MultiAction
+
+getPriority()
 - Method in class org.apache.hadoop.hbase.client.OperationWithAttributes
+
+getPriority()
 - Method in class org.apache.hadoop.hbase.client.RegionServerCallable
+
 getPriority()
 - Method in interface org.apache.hadoop.hbase.client.TableDescriptor
 
 getPriority()
 - Method in class org.apache.hadoop.hbase.client.TableDescriptorBuilder.ModifyableTableDescriptor
@@ -53729,20 +53741,21 @@
 
 HColumnDescriptor(String)
 - Constructor for class org.apache.hadoop.hbase.HColumnDescriptor
 
-Deprecated.
-Construct a column descriptor specifying only the family 
name
- The other attributes are defaulted.
+Deprecated.
+use ColumnFamilyDescriptorBuilder.of(String)
+
 
 HColumnDescriptor(byte[])
 - Constructor for class org.apache.hadoop.hbase.HColumnDescriptor
 
-Deprecated.
-Construct a column descriptor specifying only the family 
name
- The other attributes are defaulted.
+Deprecated.
+use ColumnFamilyDescriptorBuilder.of(byte[])
+
 
 HColumnDescriptor(HColumnDescriptor)
 - Constructor for class org.apache.hadoop.hbase.HColumnDescriptor
 
-Deprecated.
-Constructor.
+Deprecated.
+use ColumnFamilyDescriptorBuilder.copy(ColumnFamilyDescriptor)
+
 
 HColumnDescriptor(HColumnDescriptor,
 boolean) - Constructor for class org.apache.hadoop.hbase.HColumnDescriptor
 
@@ -72247,7 +72260,7 @@
 Callable that handles the multi method call 
going against a single
  regionserver; i.e.
 
-MultiServerCallable(ClusterConnection,
 TableName, ServerName, MultiAction, RpcController, int, 
RetryingTimeTracker) - Constructor for class 
org.apache.hadoop.hbase.client.MultiServerCallable
+MultiServerCallable(ClusterConnection,
 TableName, ServerName, MultiAction, RpcController, int, RetryingTimeTracker, 
int) - Constructor for class org.apache.hadoop.hbase.client.MultiServerCallable
 
 multiServersPerHost
 - Variable in class org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster
 
@@ -74229,7 +74242,7 @@
 
 Implementations make an rpc call against a RegionService 
via a protobuf Service.
 
-NoncedRegionServerCallable(Connection,
 TableName, byte[], HBaseRpcController) - Constructor for class 
org.apache.hadoop.hbase.client.NoncedRegionServerCallable
+NoncedRegionServerCallable(Connection,
 TableName, byte[], HBaseRpcController, int) - Constructor for class 
org.apache.hadoop.hbase.client.NoncedRegionServerCallable
 
 nonceGenerator
 - Variable in class org.apache.hadoop.hbase.client.AsyncConnectionImpl
 
@@ -74400,10 +74413,7 @@
 noReadahead
 - Variable in class org.apache.hadoop.hbase.regionserver.HStoreFile
 
 NORMAL_QOS 
- Static variable in class org.apache.hadoop.hbase.HConstants
-
-QOS attributes: 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ca5b0275/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html
index 2ca99ba..2227828 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.html
@@ -123,7 +123,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.LimitedPrivate(value="Tools")
-public class HRegionServer
+public class HRegionServer
 extends HasThread
 implements RegionServerServices, LastSequenceId, 
ConfigurationObserver
 HRegionServer makes a set of HRegions available to clients. 
It checks in with
@@ -1534,7 +1534,7 @@ implements 
 
 INIT_PAUSE_TIME_MS
-private static finalint INIT_PAUSE_TIME_MS
+private static finalint INIT_PAUSE_TIME_MS
 
 See Also:
 Constant
 Field Values
@@ -1547,7 +1547,7 @@ implements 
 
 REGION_LOCK_AWAIT_TIME_SEC
-public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String REGION_LOCK_AWAIT_TIME_SEC
+public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String REGION_LOCK_AWAIT_TIME_SEC
 
 See Also:
 Constant
 Field Values
@@ -1560,7 +1560,7 @@ implements 
 
 DEFAULT_REGION_LOCK_AWAIT_TIME_SEC
-public static finalint DEFAULT_REGION_LOCK_AWAIT_TIME_SEC
+public static finalint DEFAULT_REGION_LOCK_AWAIT_TIME_SEC
 
 See Also:
 Constant
 Field Values
@@ -1573,7 +1573,7 @@ implements 
 
 LOG
-private static finalorg.apache.commons.logging.Log LOG
+private static finalorg.apache.commons.logging.Log LOG
 
 
 
@@ -1582,7 +1582,7 @@ implements 
 
 TEST_SKIP_REPORTING_TRANSITION
-public staticboolean TEST_SKIP_REPORTING_TRANSITION
+public staticboolean TEST_SKIP_REPORTING_TRANSITION
 For testing only!  Set to true to skip notifying region 
assignment to master .
 
 
@@ -1592,7 +1592,7 @@ implements 
 
 OPEN
-protected static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String OPEN
+protected static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String OPEN
 
 See Also:
 Constant
 Field Values
@@ -1605,7 +1605,7 @@ implements 
 
 CLOSE
-protected static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String CLOSE
+protected static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String CLOSE
 
 See Also:
 Constant
 Field Values
@@ -1618,7 +1618,7 @@ implements 
 
 regionsInTransitionInRS
-protected finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true;
 title="class or interface in 
java.util.concurrent">ConcurrentMapbyte[],http://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean regionsInTransitionInRS
+protected finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true;
 title="class or interface in 
java.util.concurrent">ConcurrentMapbyte[],http://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
 title="class or interface in java.lang">Boolean regionsInTransitionInRS
 
 
 
@@ -1627,7 +1627,7 @@ implements 
 
 cacheFlusher
-protectedMemStoreFlusher cacheFlusher
+protectedMemStoreFlusher cacheFlusher
 
 
 
@@ -1636,7 +1636,7 @@ implements 
 
 hMemManager
-protectedHeapMemoryManager hMemManager
+protectedHeapMemoryManager hMemManager
 
 
 
@@ -1645,7 +1645,7 @@ implements 
 
 initLatch
-protectedhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CountDownLatch.html?is-external=true;
 title="class or interface in java.util.concurrent">CountDownLatch initLatch
+protectedhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CountDownLatch.html?is-external=true;
 title="class or interface in java.util.concurrent">CountDownLatch initLatch
 
 
 
@@ -1654,7 +1654,7 @@ implements 
 
 clusterConnection
-protectedClusterConnection clusterConnection
+protectedClusterConnection clusterConnection
 Cluster connection to be shared by services.
  Initialized at server startup and closed when server shuts down.
  Clients must never close it explicitly.
@@ -1666,7 +1666,7 @@ implements 
 
 metaTableLocator
-protectedMetaTableLocator metaTableLocator
+protectedMetaTableLocator metaTableLocator
 
 
 
@@ -1675,7 +1675,7 @@ implements 
 
 recoveringRegionWatcher
-privateRecoveringRegionWatcher 
recoveringRegionWatcher
+privateRecoveringRegionWatcher 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-19 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/9eba7fcf/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html
index 504e470..38667c0 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html
@@ -2866,5375 +2866,5371 @@
 2858checkResources();
 2859
startRegionOperation(Operation.DELETE);
 2860try {
-2861  delete.getRow();
-2862  // All edits for the given row 
(across all column families) must happen atomically.
-2863  doBatchMutate(delete);
-2864} finally {
-2865  
closeRegionOperation(Operation.DELETE);
-2866}
-2867  }
-2868
-2869  /**
-2870   * Row needed by below method.
-2871   */
-2872  private static final byte [] 
FOR_UNIT_TESTS_ONLY = Bytes.toBytes("ForUnitTestsOnly");
-2873
-2874  /**
-2875   * This is used only by unit tests. 
Not required to be a public API.
-2876   * @param familyMap map of family to 
edits for the given family.
-2877   * @throws IOException
-2878   */
-2879  void delete(NavigableMapbyte[], 
ListCell familyMap,
-2880  Durability durability) throws 
IOException {
-2881Delete delete = new 
Delete(FOR_UNIT_TESTS_ONLY);
-2882
delete.setFamilyCellMap(familyMap);
-2883delete.setDurability(durability);
-2884doBatchMutate(delete);
-2885  }
-2886
-2887  @Override
-2888  public void 
prepareDeleteTimestamps(Mutation mutation, Mapbyte[], ListCell 
familyMap,
-2889  byte[] byteNow) throws IOException 
{
-2890for (Map.Entrybyte[], 
ListCell e : familyMap.entrySet()) {
-2891
-2892  byte[] family = e.getKey();
-2893  ListCell cells = 
e.getValue();
-2894  assert cells instanceof 
RandomAccess;
-2895
-2896  Mapbyte[], Integer kvCount 
= new TreeMap(Bytes.BYTES_COMPARATOR);
-2897  int listSize = cells.size();
-2898  for (int i=0; i  listSize; 
i++) {
-2899Cell cell = cells.get(i);
-2900//  Check if time is LATEST, 
change to time of most recent addition if so
-2901//  This is expensive.
-2902if (cell.getTimestamp() == 
HConstants.LATEST_TIMESTAMP  CellUtil.isDeleteType(cell)) {
-2903  byte[] qual = 
CellUtil.cloneQualifier(cell);
-2904  if (qual == null) qual = 
HConstants.EMPTY_BYTE_ARRAY;
-2905
-2906  Integer count = 
kvCount.get(qual);
-2907  if (count == null) {
-2908kvCount.put(qual, 1);
-2909  } else {
-2910kvCount.put(qual, count + 
1);
-2911  }
-2912  count = kvCount.get(qual);
-2913
-2914  Get get = new 
Get(CellUtil.cloneRow(cell));
-2915  get.setMaxVersions(count);
-2916  get.addColumn(family, qual);
-2917  if (coprocessorHost != null) 
{
-2918if 
(!coprocessorHost.prePrepareTimeStampForDeleteVersion(mutation, cell,
-2919byteNow, get)) {
-2920  
updateDeleteLatestVersionTimeStamp(cell, get, count, byteNow);
-2921}
-2922  } else {
-2923
updateDeleteLatestVersionTimeStamp(cell, get, count, byteNow);
-2924  }
-2925} else {
-2926  
CellUtil.updateLatestStamp(cell, byteNow, 0);
-2927}
-2928  }
-2929}
-2930  }
-2931
-2932  void 
updateDeleteLatestVersionTimeStamp(Cell cell, Get get, int count, byte[] 
byteNow)
-2933  throws IOException {
-2934ListCell result = get(get, 
false);
-2935
-2936if (result.size()  count) {
-2937  // Nothing to delete
-2938  CellUtil.updateLatestStamp(cell, 
byteNow, 0);
-2939  return;
-2940}
-2941if (result.size()  count) {
-2942  throw new 
RuntimeException("Unexpected size: " + result.size());
-2943}
-2944Cell getCell = result.get(count - 
1);
-2945CellUtil.setTimestamp(cell, 
getCell.getTimestamp());
-2946  }
-2947
-2948  @Override
-2949  public void put(Put put) throws 
IOException {
-2950checkReadOnly();
-2951
-2952// Do a rough check that we have 
resources to accept a write.  The check is
-2953// 'rough' in that between the 
resource check and the call to obtain a
-2954// read lock, resources may run out. 
 For now, the thought is that this
-2955// will be extremely rare; we'll 
deal with it when it happens.
-2956checkResources();
-2957
startRegionOperation(Operation.PUT);
-2958try {
-2959  // All edits for the given row 
(across all column families) must happen atomically.
-2960  doBatchMutate(put);
-2961} finally {
-2962  
closeRegionOperation(Operation.PUT);
-2963}
-2964  }
-2965
-2966  /**
-2967   * Struct-like class that tracks the 
progress of a batch operation,
-2968   * accumulating status 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-15 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/17128d27/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html
--
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html
index 2fdac6a..def5ec9 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html
@@ -1192,686 +1192,686 @@
 1184   *
 1185   * @param regionName region to split
 1186   * @throws IOException if a remote or 
network exception occurs
-1187   */
-1188  void splitRegion(final byte[] 
regionName) throws IOException;
-1189
-1190  /**
-1191   * Split a table. Asynchronous 
operation.
-1192   *
-1193   * @param tableName table to split
-1194   * @param splitPoint the explicit 
position to split on
-1195   * @throws IOException if a remote or 
network exception occurs
-1196   */
-1197  void split(final TableName tableName, 
final byte[] splitPoint)
-1198throws IOException;
-1199
-1200  /**
-1201   * Split an individual region. 
Asynchronous operation.
-1202   *
-1203   * @param regionName region to split
-1204   * @param splitPoint the explicit 
position to split on
-1205   * @throws IOException if a remote or 
network exception occurs
-1206   */
-1207  void splitRegion(final byte[] 
regionName, final byte[] splitPoint)
-1208throws IOException;
-1209
-1210  /**
-1211   * Modify an existing table, more IRB 
friendly version.
-1212   *
-1213   * @param tableName name of table.
-1214   * @param htd modified description of 
the table
-1215   * @throws IOException if a remote or 
network exception occurs
-1216   * @deprecated since 2.0 version and 
will be removed in 3.0 version.
-1217   * use {@link 
#modifyTable(TableDescriptor)}
-1218   */
-1219  @Deprecated
-1220  void modifyTable(final TableName 
tableName, final HTableDescriptor htd)
-1221  throws IOException;
+1187   * @deprecated Since 2.0. Will be 
removed in 3.0. Use
+1188   * {@link 
#splitRegionAsync(byte[], byte[])} instead.
+1189   */
+1190  void splitRegion(final byte[] 
regionName) throws IOException;
+1191
+1192  /**
+1193   * Split a table. Asynchronous 
operation.
+1194   *
+1195   * @param tableName table to split
+1196   * @param splitPoint the explicit 
position to split on
+1197   * @throws IOException if a remote or 
network exception occurs
+1198   */
+1199  void split(final TableName tableName, 
final byte[] splitPoint)
+1200throws IOException;
+1201
+1202  /**
+1203   * Split an individual region. 
Asynchronous operation.
+1204   *
+1205   * @param regionName region to split
+1206   * @param splitPoint the explicit 
position to split on
+1207   * @throws IOException if a remote or 
network exception occurs
+1208   * @deprecated Since 2.0. Will be 
removed in 3.0. Use
+1209   * {@link 
#splitRegionAsync(byte[], byte[])} instead.
+1210   */
+1211  void splitRegion(final byte[] 
regionName, final byte[] splitPoint)
+1212throws IOException;
+1213
+1214  /**
+1215   * Split an individual region. 
Asynchronous operation.
+1216   * @param regionName region to split
+1217   * @param splitPoint the explicit 
position to split on
+1218   * @throws IOException if a remote or 
network exception occurs
+1219   */
+1220  FutureVoid 
splitRegionAsync(byte[] regionName, byte[] splitPoint)
+1221throws IOException;
 1222
 1223  /**
 1224   * Modify an existing table, more IRB 
friendly version.
 1225   *
-1226   * @param td modified description of 
the table
-1227   * @throws IOException if a remote or 
network exception occurs
-1228   */
-1229  void modifyTable(final TableDescriptor 
td) throws IOException;
-1230
-1231  /**
-1232   * Modify an existing table, more IRB 
friendly version. Asynchronous operation.  This means that
-1233   * it may be a while before your 
schema change is updated across all of the table.
-1234   * You can use Future.get(long, 
TimeUnit) to wait on the operation to complete.
-1235   * It may throw ExecutionException if 
there was an error while executing the operation
-1236   * or TimeoutException in case the 
wait timeout was not long enough to allow the
-1237   * operation to complete.
+1226   * @param tableName name of table.
+1227   * @param htd modified description of 
the table
+1228   * @throws IOException if a remote or 
network exception occurs
+1229   * @deprecated since 2.0 version and 
will be removed in 3.0 version.
+1230   * use {@link 
#modifyTable(TableDescriptor)}
+1231   */
+1232  @Deprecated
+1233  void modifyTable(final TableName 
tableName, final HTableDescriptor htd)
+1234  throws IOException;
+1235
+1236  /**
+1237   * Modify an existing table, more IRB 
friendly version.
 1238   *
-1239   * @param tableName name of table.
-1240   * @param htd modified description of 
the table
-1241   * @throws IOException if a remote or 
network 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-12 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2777c693/devapidocs/src-html/org/apache/hadoop/hbase/ipc/AbstractRpcClient.BlockingRpcChannelImplementation.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/AbstractRpcClient.BlockingRpcChannelImplementation.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/AbstractRpcClient.BlockingRpcChannelImplementation.html
index 30d5dd8..beacb44 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/AbstractRpcClient.BlockingRpcChannelImplementation.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/AbstractRpcClient.BlockingRpcChannelImplementation.html
@@ -109,497 +109,499 @@
 101  private static final 
ScheduledExecutorService IDLE_CONN_SWEEPER = Executors
 102  .newScheduledThreadPool(1, 
Threads.newDaemonThreadFactory("Idle-Rpc-Conn-Sweeper"));
 103
-104  protected final static MapKind, 
TokenSelector? extends TokenIdentifier TOKEN_HANDLERS = new 
HashMap();
-105
-106  static {
-107
TOKEN_HANDLERS.put(Kind.HBASE_AUTH_TOKEN, new AuthenticationTokenSelector());
-108  }
-109
-110  protected boolean running = true; // if 
client runs
+104  
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="MS_MUTABLE_COLLECTION_PKGPROTECT",
+105  justification="the rest of the 
system which live in the different package can use")
+106  protected final static MapKind, 
TokenSelector? extends TokenIdentifier TOKEN_HANDLERS = new 
HashMap();
+107
+108  static {
+109
TOKEN_HANDLERS.put(Kind.HBASE_AUTH_TOKEN, new AuthenticationTokenSelector());
+110  }
 111
-112  protected final Configuration conf;
-113  protected final String clusterId;
-114  protected final SocketAddress 
localAddr;
-115  protected final MetricsConnection 
metrics;
-116
-117  protected final UserProvider 
userProvider;
-118  protected final CellBlockBuilder 
cellBlockBuilder;
-119
-120  protected final int 
minIdleTimeBeforeClose; // if the connection is idle for more than this
-121  // time (in ms), it will be closed at 
any moment.
-122  protected final int maxRetries; // the 
max. no. of retries for socket connections
-123  protected final long failureSleep; // 
Time to sleep before retry on failure.
-124  protected final boolean tcpNoDelay; // 
if T then disable Nagle's Algorithm
-125  protected final boolean tcpKeepAlive; 
// if T then use keepalives
-126  protected final Codec codec;
-127  protected final CompressionCodec 
compressor;
-128  protected final boolean 
fallbackAllowed;
-129
-130  protected final FailedServers 
failedServers;
+112  protected boolean running = true; // if 
client runs
+113
+114  protected final Configuration conf;
+115  protected final String clusterId;
+116  protected final SocketAddress 
localAddr;
+117  protected final MetricsConnection 
metrics;
+118
+119  protected final UserProvider 
userProvider;
+120  protected final CellBlockBuilder 
cellBlockBuilder;
+121
+122  protected final int 
minIdleTimeBeforeClose; // if the connection is idle for more than this
+123  // time (in ms), it will be closed at 
any moment.
+124  protected final int maxRetries; // the 
max. no. of retries for socket connections
+125  protected final long failureSleep; // 
Time to sleep before retry on failure.
+126  protected final boolean tcpNoDelay; // 
if T then disable Nagle's Algorithm
+127  protected final boolean tcpKeepAlive; 
// if T then use keepalives
+128  protected final Codec codec;
+129  protected final CompressionCodec 
compressor;
+130  protected final boolean 
fallbackAllowed;
 131
-132  protected final int connectTO;
-133  protected final int readTO;
-134  protected final int writeTO;
-135
-136  protected final 
PoolMapConnectionId, T connections;
+132  protected final FailedServers 
failedServers;
+133
+134  protected final int connectTO;
+135  protected final int readTO;
+136  protected final int writeTO;
 137
-138  private final AtomicInteger callIdCnt = 
new AtomicInteger(0);
+138  protected final 
PoolMapConnectionId, T connections;
 139
-140  private final ScheduledFuture? 
cleanupIdleConnectionTask;
+140  private final AtomicInteger callIdCnt = 
new AtomicInteger(0);
 141
-142  private int 
maxConcurrentCallsPerServer;
+142  private final ScheduledFuture? 
cleanupIdleConnectionTask;
 143
-144  private static final 
LoadingCacheInetSocketAddress, AtomicInteger concurrentCounterCache =
-145  
CacheBuilder.newBuilder().expireAfterAccess(1, TimeUnit.HOURS).
-146  build(new 
CacheLoaderInetSocketAddress, AtomicInteger() {
-147@Override public 
AtomicInteger load(InetSocketAddress key) throws Exception {
-148  return new 
AtomicInteger(0);
-149}
-150  });
-151
-152  /**
-153   * Construct an IPC client for the 
cluster codeclusterId/code
-154   * @param conf configuration
-155   * @param clusterId the cluster id
-156   * @param localAddr client socket bind 
address.
-157   * @param metrics the 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-11 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/90c7dfe4/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html
index 75db22d..99a09f9 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html
@@ -37,2710 +37,2816 @@
 029import java.util.List;
 030import java.util.Map;
 031import java.util.Optional;
-032import 
java.util.concurrent.CompletableFuture;
-033import java.util.concurrent.TimeUnit;
-034import 
java.util.concurrent.atomic.AtomicReference;
-035import java.util.function.BiConsumer;
-036import java.util.regex.Pattern;
-037import java.util.stream.Collectors;
-038
-039import 
com.google.common.annotations.VisibleForTesting;
-040
-041import io.netty.util.Timeout;
-042import io.netty.util.TimerTask;
-043
-044import java.util.stream.Stream;
-045
-046import org.apache.commons.io.IOUtils;
-047import org.apache.commons.logging.Log;
-048import 
org.apache.commons.logging.LogFactory;
-049import 
org.apache.hadoop.hbase.ClusterStatus;
-050import 
org.apache.hadoop.hbase.HRegionInfo;
-051import 
org.apache.hadoop.hbase.HRegionLocation;
-052import 
org.apache.hadoop.hbase.MetaTableAccessor;
-053import 
org.apache.hadoop.hbase.MetaTableAccessor.QueryType;
-054import 
org.apache.hadoop.hbase.NotServingRegionException;
-055import 
org.apache.hadoop.hbase.ProcedureInfo;
-056import 
org.apache.hadoop.hbase.RegionLoad;
-057import 
org.apache.hadoop.hbase.RegionLocations;
-058import 
org.apache.hadoop.hbase.ServerName;
-059import 
org.apache.hadoop.hbase.NamespaceDescriptor;
-060import 
org.apache.hadoop.hbase.HConstants;
-061import 
org.apache.hadoop.hbase.TableExistsException;
-062import 
org.apache.hadoop.hbase.TableName;
-063import 
org.apache.hadoop.hbase.AsyncMetaTableAccessor;
-064import 
org.apache.hadoop.hbase.TableNotDisabledException;
-065import 
org.apache.hadoop.hbase.TableNotEnabledException;
-066import 
org.apache.hadoop.hbase.TableNotFoundException;
-067import 
org.apache.hadoop.hbase.UnknownRegionException;
-068import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-069import 
org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder;
-070import 
org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder;
-071import 
org.apache.hadoop.hbase.client.Scan.ReadType;
-072import 
org.apache.hadoop.hbase.client.replication.ReplicationSerDeHelper;
-073import 
org.apache.hadoop.hbase.client.replication.TableCFs;
-074import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-075import 
org.apache.hadoop.hbase.ipc.HBaseRpcController;
-076import 
org.apache.hadoop.hbase.quotas.QuotaFilter;
-077import 
org.apache.hadoop.hbase.quotas.QuotaSettings;
-078import 
org.apache.hadoop.hbase.quotas.QuotaTableUtil;
-079import 
org.apache.hadoop.hbase.replication.ReplicationException;
-080import 
org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
-081import 
org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
-082import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback;
-083import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-084import 
org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
-085import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
-086import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest;
-087import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse;
-088import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest;
-089import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse;
-090import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest;
-091import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionResponse;
-092import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionRequest;
-093import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionResponse;
-094import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest;
-095import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse;
-096import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequest;
-097import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse;
-098import 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0821e51a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteNamespaceProcedureBiConsumer.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteNamespaceProcedureBiConsumer.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteNamespaceProcedureBiConsumer.html
index 71844ce..75db22d 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteNamespaceProcedureBiConsumer.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteNamespaceProcedureBiConsumer.html
@@ -105,2564 +105,2642 @@
 097import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse;
 098import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionRequest;
 099import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionResponse;
-100import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription;
-101import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema;
-102import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType;
-103import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest;
-104import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse;
-105import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest;
-106import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse;
-107import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest;
-108import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse;
-109import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest;
-110import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse;
-111import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest;
-112import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse;
-113import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest;
-114import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse;
-115import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest;
-116import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse;
-117import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest;
-118import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse;
-119import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest;
-120import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse;
-121import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest;
-122import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse;
-123import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest;
-124import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse;
-125import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest;
-126import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse;
-127import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest;
-128import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse;
-129import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest;
-130import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse;
-131import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest;
-132import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse;
-133import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest;
-134import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse;
-135import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest;
-136import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse;
-137import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest;
-138import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse;
-139import 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-08 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/9fb0764b/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html
--
diff --git 
a/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html 
b/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html
index 2def063..4e1f2ed 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html
@@ -62,854 +62,860 @@
 054import 
org.apache.hadoop.hbase.client.RowMutations;
 055import 
org.apache.hadoop.hbase.client.Scan;
 056import 
org.apache.hadoop.hbase.client.Table;
-057import 
org.apache.hadoop.hbase.client.coprocessor.Batch;
-058import 
org.apache.hadoop.hbase.client.coprocessor.Batch.Callback;
-059import 
org.apache.hadoop.hbase.client.metrics.ScanMetrics;
-060import 
org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
-061import 
org.apache.hadoop.hbase.io.TimeRange;
-062import 
org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
-063import 
org.apache.hadoop.hbase.rest.Constants;
-064import 
org.apache.hadoop.hbase.rest.model.CellModel;
-065import 
org.apache.hadoop.hbase.rest.model.CellSetModel;
-066import 
org.apache.hadoop.hbase.rest.model.RowModel;
-067import 
org.apache.hadoop.hbase.rest.model.ScannerModel;
-068import 
org.apache.hadoop.hbase.rest.model.TableSchemaModel;
-069import 
org.apache.hadoop.hbase.util.Bytes;
-070import 
org.apache.hadoop.util.StringUtils;
-071
-072import com.google.protobuf.Descriptors;
-073import com.google.protobuf.Message;
-074import com.google.protobuf.Service;
-075import 
com.google.protobuf.ServiceException;
-076
-077/**
-078 * HTable interface to remote tables 
accessed via REST gateway
-079 */
-080@InterfaceAudience.Public
-081public class RemoteHTable implements 
Table {
-082
-083  private static final Log LOG = 
LogFactory.getLog(RemoteHTable.class);
-084
-085  final Client client;
-086  final Configuration conf;
-087  final byte[] name;
-088  final int maxRetries;
-089  final long sleepTime;
-090
-091  @SuppressWarnings("rawtypes")
-092  protected String buildRowSpec(final 
byte[] row, final Map familyMap,
-093  final long startTime, final long 
endTime, final int maxVersions) {
-094StringBuffer sb = new 
StringBuffer();
-095sb.append('/');
-096sb.append(Bytes.toString(name));
-097sb.append('/');
-098sb.append(toURLEncodedBytes(row));
-099Set families = 
familyMap.entrySet();
-100if (families != null) {
-101  Iterator i = 
familyMap.entrySet().iterator();
-102  sb.append('/');
-103  while (i.hasNext()) {
-104Map.Entry e = 
(Map.Entry)i.next();
-105Collection quals = 
(Collection)e.getValue();
-106if (quals == null || 
quals.isEmpty()) {
-107  // this is an unqualified 
family. append the family name and NO ':'
-108  
sb.append(toURLEncodedBytes((byte[])e.getKey()));
-109} else {
-110  Iterator ii = 
quals.iterator();
-111  while (ii.hasNext()) {
-112
sb.append(toURLEncodedBytes((byte[])e.getKey()));
-113sb.append(':');
-114Object o = ii.next();
-115// Puts use byte[] but 
Deletes use KeyValue
-116if (o instanceof byte[]) {
-117  
sb.append(toURLEncodedBytes((byte[])o));
-118} else if (o instanceof 
KeyValue) {
-119  
sb.append(toURLEncodedBytes(CellUtil.cloneQualifier((KeyValue)o)));
-120} else {
-121  throw new 
RuntimeException("object type not handled");
-122}
-123if (ii.hasNext()) {
-124  sb.append(',');
-125}
-126  }
-127}
-128if (i.hasNext()) {
-129  sb.append(',');
-130}
-131  }
-132}
-133if (startTime = 0  
endTime != Long.MAX_VALUE) {
-134  sb.append('/');
-135  sb.append(startTime);
-136  if (startTime != endTime) {
-137sb.append(',');
-138sb.append(endTime);
-139  }
-140} else if (endTime != Long.MAX_VALUE) 
{
-141  sb.append('/');
-142  sb.append(endTime);
-143}
-144if (maxVersions  1) {
-145  sb.append("?v=");
-146  sb.append(maxVersions);
-147}
-148return sb.toString();
-149  }
-150
-151  protected String 
buildMultiRowSpec(final byte[][] rows, int maxVersions) {
-152StringBuilder sb = new 
StringBuilder();
-153sb.append('/');
-154sb.append(Bytes.toString(name));
-155sb.append("/multiget/");
-156if (rows == null || rows.length == 0) 
{
-157  return sb.toString();
-158}
-159sb.append("?");
-160for(int i=0; irows.length; i++) 
{
-161  byte[] rk = rows[i];
-162  if (i != 0) {
-163sb.append('');
-164  }
-165  sb.append("row=");
-166  sb.append(toURLEncodedBytes(rk));
-167}
-168sb.append("v=");
-169sb.append(maxVersions);
-170
-171  

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-07 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b3b50f22/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html
index 249d4a0..7369fdf 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.Writer.State.html
@@ -65,12 +65,12 @@
 057import 
com.google.common.base.Preconditions;
 058
 059/**
-060 * Reads {@link HFile} version 2 blocks 
to HFiles and via {@link Cacheable} Interface to caches.
-061 * Version 2 was introduced in 
hbase-0.92.0. No longer has support for version 1 blocks since
-062 * hbase-1.3.0.
-063 *
-064 * pVersion 1 was the original 
file block. Version 2 was introduced when we changed the hbase file
-065 * format to support multi-level block 
indexes and compound bloom filters (HBASE-3857).
+060 * Cacheable Blocks of an {@link HFile} 
version 2 file.
+061 * Version 2 was introduced in 
hbase-0.92.0.
+062 *
+063 * pVersion 1 was the original 
file block. Version 2 was introduced when we changed the hbase file
+064 * format to support multi-level block 
indexes and compound bloom filters (HBASE-3857). Support
+065 * for Version 1 was removed in 
hbase-1.3.0.
 066 *
 067 * h3HFileBlock: Version 
2/h3
 068 * In version 2, a block is structured as 
follows:
@@ -120,582 +120,582 @@
 112public class HFileBlock implements 
Cacheable {
 113  private static final Log LOG = 
LogFactory.getLog(HFileBlock.class);
 114
-115  /** Type of block. Header field 0. */
-116  private BlockType blockType;
-117
-118  /**
-119   * Size on disk excluding header, 
including checksum. Header field 1.
-120   * @see Writer#putHeader(byte[], int, 
int, int, int)
-121   */
-122  private int onDiskSizeWithoutHeader;
-123
-124  /**
-125   * Size of pure data. Does not include 
header or checksums. Header field 2.
-126   * @see Writer#putHeader(byte[], int, 
int, int, int)
-127   */
-128  private int 
uncompressedSizeWithoutHeader;
-129
-130  /**
-131   * The offset of the previous block on 
disk. Header field 3.
-132   * @see Writer#putHeader(byte[], int, 
int, int, int)
-133   */
-134  private long prevBlockOffset;
-135
-136  /**
-137   * Size on disk of header + data. 
Excludes checksum. Header field 6,
-138   * OR calculated from {@link 
#onDiskSizeWithoutHeader} when using HDFS checksum.
-139   * @see Writer#putHeader(byte[], int, 
int, int, int)
-140   */
-141  private int onDiskDataSizeWithHeader;
-142
-143
-144  /**
-145   * The in-memory representation of the 
hfile block. Can be on or offheap. Can be backed by
-146   * a single ByteBuffer or by many. Make 
no assumptions.
-147   *
-148   * pBe careful reading from 
this codebuf/code. Duplicate and work on the duplicate or if
-149   * not, be sure to reset position and 
limit else trouble down the road.
-150   *
-151   * pTODO: Make this read-only 
once made.
-152   *
-153   * pWe are using the ByteBuff 
type. ByteBuffer is not extensible yet we need to be able to have
-154   * a ByteBuffer-like API across 
multiple ByteBuffers reading from a cache such as BucketCache.
-155   * So, we have this ByteBuff type. 
Unfortunately, it is spread all about HFileBlock. Would be
-156   * good if could be confined to 
cache-use only but hard-to-do.
-157   */
-158  private ByteBuff buf;
-159
-160  /** Meta data that holds meta 
information on the hfileblock.
-161   */
-162  private HFileContext fileContext;
-163
-164  /**
-165   * The offset of this block in the 
file. Populated by the reader for
-166   * convenience of access. This offset 
is not part of the block header.
-167   */
-168  private long offset = UNSET;
-169
-170  private MemoryType memType = 
MemoryType.EXCLUSIVE;
-171
-172  /**
-173   * The on-disk size of the next block, 
including the header and checksums if present, obtained by
-174   * peeking into the first {@link 
HConstants#HFILEBLOCK_HEADER_SIZE} bytes of the next block's
-175   * header, or UNSET if unknown.
-176   *
-177   * Blocks try to carry the size of the 
next block to read in this data member. They will even have
-178   * this value when served from cache. 
Could save a seek in the case where we are iterating through
-179   * a file and some of the blocks come 
from cache. If from cache, then having this info to hand
-180   * will save us doing a seek to read 
the header so we can read the body of a block.
-181   * TODO: see how effective this is at 
saving seeks.
-182   */
-183  private int nextBlockOnDiskSize = 
UNSET;
-184
-185  /**
-186   * On a checksum failure, do these many 
succeeding read requests using hdfs checksums before
-187   * auto-reenabling hbase checksum 
verification.
-188   */
-189  static final int 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-05 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ca9f6925/devapidocs/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.DisableTableProcedureBiConsumer.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.DisableTableProcedureBiConsumer.html
 
b/devapidocs/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.DisableTableProcedureBiConsumer.html
deleted file mode 100644
index 5e6de54..000
--- 
a/devapidocs/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.DisableTableProcedureBiConsumer.html
+++ /dev/null
@@ -1,339 +0,0 @@
-http://www.w3.org/TR/html4/loose.dtd;>
-
-
-
-
-
-AsyncHBaseAdmin.DisableTableProcedureBiConsumer (Apache HBase 
3.0.0-SNAPSHOT API)
-
-
-
-
-
-var methods = {"i0":10};
-var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
-var altColor = "altColor";
-var rowColor = "rowColor";
-var tableTab = "tableTab";
-var activeTableTab = "activeTableTab";
-
-
-JavaScript is disabled on your browser.
-
-
-
-
-
-Skip navigation links
-
-
-
-
-Overview
-Package
-Class
-Use
-Tree
-Deprecated
-Index
-Help
-
-
-
-
-PrevClass
-NextClass
-
-
-Frames
-NoFrames
-
-
-AllClasses
-
-
-
-
-
-
-
-Summary:
-Nested|
-Field|
-Constr|
-Method
-
-
-Detail:
-Field|
-Constr|
-Method
-
-
-
-
-
-
-
-
-org.apache.hadoop.hbase.client
-Class AsyncHBaseAdmin.DisableTableProcedureBiConsumer
-
-
-
-http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">java.lang.Object
-
-
-org.apache.hadoop.hbase.client.AsyncHBaseAdmin.ProcedureBiConsumer
-
-
-org.apache.hadoop.hbase.client.AsyncHBaseAdmin.TableProcedureBiConsumer
-
-
-org.apache.hadoop.hbase.client.AsyncHBaseAdmin.DisableTableProcedureBiConsumer
-
-
-
-
-
-
-
-
-
-
-
-All Implemented Interfaces:
-http://docs.oracle.com/javase/8/docs/api/java/util/function/BiConsumer.html?is-external=true;
 title="class or interface in java.util.function">BiConsumerhttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void,http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
 title="class or interface in java.lang">Throwable
-
-
-Enclosing class:
-AsyncHBaseAdmin
-
-
-
-private class AsyncHBaseAdmin.DisableTableProcedureBiConsumer
-extends AsyncHBaseAdmin.TableProcedureBiConsumer
-
-
-
-
-
-
-
-
-
-
-
-Field Summary
-
-
-
-
-Fields inherited from classorg.apache.hadoop.hbase.client.AsyncHBaseAdmin.TableProcedureBiConsumer
-tableName
-
-
-
-
-
-Fields inherited from classorg.apache.hadoop.hbase.client.AsyncHBaseAdmin.ProcedureBiConsumer
-admin
-
-
-
-
-
-
-
-
-Constructor Summary
-
-Constructors
-
-Constructor and Description
-
-
-DisableTableProcedureBiConsumer(AsyncAdminadmin,
-   TableNametableName)
-
-
-
-
-
-
-
-
-
-Method Summary
-
-All MethodsInstance MethodsConcrete Methods
-
-Modifier and Type
-Method and Description
-
-
-(package private) http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
-getOperationType()
-
-
-
-
-
-
-Methods inherited from classorg.apache.hadoop.hbase.client.AsyncHBaseAdmin.TableProcedureBiConsumer
-getDescription,
 onError,
 onFinished
-
-
-
-
-
-Methods inherited from classorg.apache.hadoop.hbase.client.AsyncHBaseAdmin.ProcedureBiConsumer
-accept
-
-
-
-
-
-Methods inherited from classjava.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
-http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--;
 title="class or interface in java.lang">clone, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-;
 title="class or interface in java.lang">equals, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--;
 title="class or interface in java.lang">finalize, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--;
 title="class or interface in java.lang">getClass, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--;
 title="class or interface in java.lang">hashCode, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--;
 title="class or interface in java.lang">notify, 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-06-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8e3b63ca/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.MasterRpcCall.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.MasterRpcCall.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.MasterRpcCall.html
index dc12c09..82506d2 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.MasterRpcCall.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.MasterRpcCall.html
@@ -54,2261 +54,2259 @@
 046import org.apache.commons.io.IOUtils;
 047import org.apache.commons.logging.Log;
 048import 
org.apache.commons.logging.LogFactory;
-049import 
org.apache.directory.api.util.OptionalComponentsMonitor;
-050import 
org.apache.hadoop.hbase.HRegionInfo;
-051import 
org.apache.hadoop.hbase.HRegionLocation;
-052import 
org.apache.hadoop.hbase.MetaTableAccessor;
-053import 
org.apache.hadoop.hbase.MetaTableAccessor.QueryType;
-054import 
org.apache.hadoop.hbase.NotServingRegionException;
-055import 
org.apache.hadoop.hbase.ProcedureInfo;
-056import 
org.apache.hadoop.hbase.RegionLocations;
-057import 
org.apache.hadoop.hbase.ServerName;
-058import 
org.apache.hadoop.hbase.NamespaceDescriptor;
-059import 
org.apache.hadoop.hbase.HConstants;
-060import 
org.apache.hadoop.hbase.TableExistsException;
-061import 
org.apache.hadoop.hbase.TableName;
-062import 
org.apache.hadoop.hbase.AsyncMetaTableAccessor;
-063import 
org.apache.hadoop.hbase.TableNotDisabledException;
-064import 
org.apache.hadoop.hbase.TableNotEnabledException;
-065import 
org.apache.hadoop.hbase.TableNotFoundException;
-066import 
org.apache.hadoop.hbase.UnknownRegionException;
-067import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-068import 
org.apache.hadoop.hbase.classification.InterfaceStability;
-069import 
org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder;
-070import 
org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder;
-071import 
org.apache.hadoop.hbase.client.Scan.ReadType;
-072import 
org.apache.hadoop.hbase.client.replication.ReplicationSerDeHelper;
-073import 
org.apache.hadoop.hbase.client.replication.TableCFs;
-074import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-075import 
org.apache.hadoop.hbase.ipc.HBaseRpcController;
-076import 
org.apache.hadoop.hbase.quotas.QuotaFilter;
-077import 
org.apache.hadoop.hbase.quotas.QuotaSettings;
-078import 
org.apache.hadoop.hbase.quotas.QuotaTableUtil;
-079import 
org.apache.hadoop.hbase.replication.ReplicationException;
-080import 
org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
-081import 
org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
-082import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback;
-083import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-084import 
org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
-085import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
-086import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest;
-087import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse;
-088import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest;
-089import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse;
-090import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest;
-091import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionResponse;
-092import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionRequest;
-093import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionResponse;
-094import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionRequest;
-095import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionResponse;
-096import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription;
-097import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema;
-098import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest;
-099import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse;
-100import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest;
-101import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse;
-102import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest;
-103import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse;
-104import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest;
-105import 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-06-26 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aecb1286/devapidocs/src-html/org/apache/hadoop/hbase/AsyncMetaTableAccessor.MetaTableRawScanResultConsumer.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/AsyncMetaTableAccessor.MetaTableRawScanResultConsumer.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/AsyncMetaTableAccessor.MetaTableRawScanResultConsumer.html
index baee8c1..a0256ee 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/AsyncMetaTableAccessor.MetaTableRawScanResultConsumer.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/AsyncMetaTableAccessor.MetaTableRawScanResultConsumer.html
@@ -30,551 +30,607 @@
 022import java.io.IOException;
 023import java.util.ArrayList;
 024import java.util.Arrays;
-025import java.util.List;
-026import java.util.Map;
-027import java.util.NavigableMap;
-028import java.util.Optional;
-029import java.util.SortedMap;
-030import 
java.util.concurrent.CompletableFuture;
-031import java.util.regex.Matcher;
-032import java.util.regex.Pattern;
-033
-034import org.apache.commons.logging.Log;
-035import 
org.apache.commons.logging.LogFactory;
-036import 
org.apache.hadoop.hbase.MetaTableAccessor.CollectingVisitor;
-037import 
org.apache.hadoop.hbase.MetaTableAccessor.QueryType;
-038import 
org.apache.hadoop.hbase.MetaTableAccessor.Visitor;
-039import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-040import 
org.apache.hadoop.hbase.client.Consistency;
-041import 
org.apache.hadoop.hbase.client.Get;
-042import 
org.apache.hadoop.hbase.client.RawAsyncTable;
-043import 
org.apache.hadoop.hbase.client.RawScanResultConsumer;
-044import 
org.apache.hadoop.hbase.client.RegionReplicaUtil;
-045import 
org.apache.hadoop.hbase.client.Result;
-046import 
org.apache.hadoop.hbase.client.Scan;
-047import 
org.apache.hadoop.hbase.client.TableState;
-048import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-049import 
org.apache.hadoop.hbase.util.Bytes;
-050import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-051import 
org.apache.hadoop.hbase.util.Pair;
-052
-053/**
-054 * The asynchronous meta table accessor. 
Used to read/write region and assignment information store
-055 * in 
codehbase:meta/code.
-056 */
-057@InterfaceAudience.Private
-058public class AsyncMetaTableAccessor {
-059
-060  private static final Log LOG = 
LogFactory.getLog(AsyncMetaTableAccessor.class);
-061
-062
-063  /** The delimiter for meta columns for 
replicaIds gt; 0 */
-064  private static final char 
META_REPLICA_ID_DELIMITER = '_';
+025import java.util.Collections;
+026import java.util.List;
+027import java.util.Map;
+028import java.util.NavigableMap;
+029import java.util.Optional;
+030import java.util.SortedMap;
+031import 
java.util.concurrent.CompletableFuture;
+032import java.util.regex.Matcher;
+033import java.util.regex.Pattern;
+034import java.util.stream.Collectors;
+035
+036import org.apache.commons.logging.Log;
+037import 
org.apache.commons.logging.LogFactory;
+038import 
org.apache.hadoop.hbase.MetaTableAccessor.CollectingVisitor;
+039import 
org.apache.hadoop.hbase.MetaTableAccessor.QueryType;
+040import 
org.apache.hadoop.hbase.MetaTableAccessor.Visitor;
+041import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
+042import 
org.apache.hadoop.hbase.client.Connection;
+043import 
org.apache.hadoop.hbase.client.Consistency;
+044import 
org.apache.hadoop.hbase.client.Get;
+045import 
org.apache.hadoop.hbase.client.RawAsyncTable;
+046import 
org.apache.hadoop.hbase.client.RawScanResultConsumer;
+047import 
org.apache.hadoop.hbase.client.RegionReplicaUtil;
+048import 
org.apache.hadoop.hbase.client.Result;
+049import 
org.apache.hadoop.hbase.client.Scan;
+050import 
org.apache.hadoop.hbase.client.TableState;
+051import 
org.apache.hadoop.hbase.client.Scan.ReadType;
+052import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
+053import 
org.apache.hadoop.hbase.util.Bytes;
+054import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
+055import 
org.apache.hadoop.hbase.util.Pair;
+056
+057/**
+058 * The asynchronous meta table accessor. 
Used to read/write region and assignment information store
+059 * in 
codehbase:meta/code.
+060 */
+061@InterfaceAudience.Private
+062public class AsyncMetaTableAccessor {
+063
+064  private static final Log LOG = 
LogFactory.getLog(AsyncMetaTableAccessor.class);
 065
-066  /** A regex for parsing server columns 
from meta. See above javadoc for meta layout */
-067  private static final Pattern 
SERVER_COLUMN_PATTERN = Pattern
-068  
.compile("^server(_[0-9a-fA-F]{4})?$");
+066
+067  /** The delimiter for meta columns for 
replicaIds gt; 0 */
+068  private static final char 
META_REPLICA_ID_DELIMITER = '_';
 069
-070  public static 
CompletableFutureBoolean tableExists(RawAsyncTable metaTable, TableName 
tableName) {
-071if 
(tableName.equals(META_TABLE_NAME)) {
-072  return 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-06-20 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a719cd00/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html
index 6de986f..c895448 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html
@@ -26,1592 +26,1693 @@
 018package 
org.apache.hadoop.hbase.master.balancer;
 019
 020import java.util.ArrayDeque;
-021import java.util.Arrays;
-022import java.util.Collection;
-023import java.util.Deque;
-024import java.util.HashMap;
-025import java.util.LinkedList;
-026import java.util.List;
-027import java.util.Map;
-028import java.util.Map.Entry;
-029import java.util.Random;
-030
-031import org.apache.commons.logging.Log;
-032import 
org.apache.commons.logging.LogFactory;
-033import 
org.apache.hadoop.conf.Configuration;
-034import 
org.apache.hadoop.hbase.ClusterStatus;
-035import 
org.apache.hadoop.hbase.HBaseInterfaceAudience;
-036import 
org.apache.hadoop.hbase.HConstants;
-037import 
org.apache.hadoop.hbase.HRegionInfo;
-038import 
org.apache.hadoop.hbase.RegionLoad;
-039import 
org.apache.hadoop.hbase.ServerLoad;
-040import 
org.apache.hadoop.hbase.ServerName;
-041import 
org.apache.hadoop.hbase.TableName;
-042import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-043import 
org.apache.hadoop.hbase.master.MasterServices;
-044import 
org.apache.hadoop.hbase.master.RegionPlan;
-045import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action;
-046import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action.Type;
-047import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.AssignRegionAction;
-048import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.MoveRegionAction;
-049import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.SwapRegionsAction;
-050import 
org.apache.hadoop.hbase.util.Bytes;
-051import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-052
-053import com.google.common.collect.Lists;
-054
-055/**
-056 * pThis is a best effort load 
balancer. Given a Cost function F(C) =gt; x It will
-057 * randomly try and mutate the cluster to 
Cprime. If F(Cprime) lt; F(C) then the
-058 * new cluster state becomes the plan. It 
includes costs functions to compute the cost of:/p
-059 * ul
-060 * liRegion Load/li
-061 * liTable Load/li
-062 * liData Locality/li
-063 * liMemstore Sizes/li
-064 * liStorefile Sizes/li
-065 * /ul
-066 *
-067 *
-068 * pEvery cost function returns a 
number between 0 and 1 inclusive; where 0 is the lowest cost
-069 * best solution, and 1 is the highest 
possible cost and the worst solution.  The computed costs are
-070 * scaled by their respective 
multipliers:/p
+021import java.util.ArrayList;
+022import java.util.Arrays;
+023import java.util.Collection;
+024import java.util.Collections;
+025import java.util.Deque;
+026import java.util.HashMap;
+027import java.util.LinkedList;
+028import java.util.List;
+029import java.util.Map;
+030import java.util.Map.Entry;
+031import java.util.Random;
+032
+033import org.apache.commons.logging.Log;
+034import 
org.apache.commons.logging.LogFactory;
+035import 
org.apache.hadoop.conf.Configuration;
+036import 
org.apache.hadoop.hbase.ClusterStatus;
+037import 
org.apache.hadoop.hbase.HBaseInterfaceAudience;
+038import 
org.apache.hadoop.hbase.HConstants;
+039import 
org.apache.hadoop.hbase.HRegionInfo;
+040import 
org.apache.hadoop.hbase.RegionLoad;
+041import 
org.apache.hadoop.hbase.ServerLoad;
+042import 
org.apache.hadoop.hbase.ServerName;
+043import 
org.apache.hadoop.hbase.TableName;
+044import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
+045import 
org.apache.hadoop.hbase.master.MasterServices;
+046import 
org.apache.hadoop.hbase.master.RegionPlan;
+047import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action;
+048import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action.Type;
+049import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.AssignRegionAction;
+050import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.LocalityType;
+051import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.MoveRegionAction;
+052import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.SwapRegionsAction;
+053import 
org.apache.hadoop.hbase.util.Bytes;
+054import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
+055
+056import com.google.common.base.Optional;
+057import 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-06-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/476c54ed/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
index 0c3fe3b..d262744 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
@@ -149,3339 +149,3348 @@
 141import 
org.apache.hadoop.hbase.quotas.MasterSpaceQuotaObserver;
 142import 
org.apache.hadoop.hbase.quotas.QuotaObserverChore;
 143import 
org.apache.hadoop.hbase.quotas.QuotaUtil;
-144import 
org.apache.hadoop.hbase.quotas.SpaceQuotaSnapshotNotifier;
-145import 
org.apache.hadoop.hbase.quotas.SpaceQuotaSnapshotNotifierFactory;
-146import 
org.apache.hadoop.hbase.regionserver.DefaultStoreEngine;
-147import 
org.apache.hadoop.hbase.regionserver.HRegionServer;
-148import 
org.apache.hadoop.hbase.regionserver.HStore;
-149import 
org.apache.hadoop.hbase.regionserver.RSRpcServices;
-150import 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
-151import 
org.apache.hadoop.hbase.regionserver.RegionSplitPolicy;
-152import 
org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy;
-153import 
org.apache.hadoop.hbase.regionserver.compactions.FIFOCompactionPolicy;
-154import 
org.apache.hadoop.hbase.replication.ReplicationException;
-155import 
org.apache.hadoop.hbase.replication.ReplicationFactory;
-156import 
org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
-157import 
org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
-158import 
org.apache.hadoop.hbase.replication.ReplicationQueuesZKImpl;
-159import 
org.apache.hadoop.hbase.replication.master.TableCFsUpdater;
-160import 
org.apache.hadoop.hbase.replication.regionserver.Replication;
-161import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-162import 
org.apache.hadoop.hbase.security.UserProvider;
-163import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-164import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState;
-165import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionServerInfo;
-166import 
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas;
-167import 
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceViolationPolicy;
-168import 
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription;
-169import 
org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos;
-170import 
org.apache.hadoop.hbase.util.Addressing;
-171import 
org.apache.hadoop.hbase.util.Bytes;
-172import 
org.apache.hadoop.hbase.util.CompressionTest;
-173import 
org.apache.hadoop.hbase.util.EncryptionTest;
-174import 
org.apache.hadoop.hbase.util.FSUtils;
-175import 
org.apache.hadoop.hbase.util.HFileArchiveUtil;
-176import 
org.apache.hadoop.hbase.util.HasThread;
-177import 
org.apache.hadoop.hbase.util.IdLock;
-178import 
org.apache.hadoop.hbase.util.ModifyRegionUtils;
-179import 
org.apache.hadoop.hbase.util.Pair;
-180import 
org.apache.hadoop.hbase.util.Threads;
-181import 
org.apache.hadoop.hbase.util.VersionInfo;
-182import 
org.apache.hadoop.hbase.util.ZKDataMigrator;
-183import 
org.apache.hadoop.hbase.zookeeper.DrainingServerTracker;
-184import 
org.apache.hadoop.hbase.zookeeper.LoadBalancerTracker;
-185import 
org.apache.hadoop.hbase.zookeeper.MasterAddressTracker;
-186import 
org.apache.hadoop.hbase.zookeeper.MasterMaintenanceModeTracker;
-187import 
org.apache.hadoop.hbase.zookeeper.RegionNormalizerTracker;
-188import 
org.apache.hadoop.hbase.zookeeper.RegionServerTracker;
-189import 
org.apache.hadoop.hbase.zookeeper.SplitOrMergeTracker;
-190import 
org.apache.hadoop.hbase.zookeeper.ZKClusterId;
-191import 
org.apache.hadoop.hbase.zookeeper.ZKUtil;
-192import 
org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher;
-193import 
org.apache.zookeeper.KeeperException;
-194import org.eclipse.jetty.server.Server;
-195import 
org.eclipse.jetty.server.ServerConnector;
-196import 
org.eclipse.jetty.servlet.ServletHolder;
-197import 
org.eclipse.jetty.webapp.WebAppContext;
-198
-199import 
com.google.common.annotations.VisibleForTesting;
-200import com.google.common.collect.Lists;
-201import com.google.common.collect.Maps;
-202import com.google.protobuf.Descriptors;
-203import com.google.protobuf.Service;
-204
-205/**
-206 * HMaster is the "master server" for 
HBase. An HBase cluster has one active
-207 * master.  If many masters are started, 
all compete.  Whichever wins goes on to
-208 * run the cluster.  All others park 
themselves in their constructor until
-209 * master or cluster shutdown or until 
the active master loses its lease in
-210 * zookeeper.  Thereafter, all 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-06-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/77a552c4/devapidocs/org/apache/hadoop/hbase/HTableDescriptor.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/HTableDescriptor.html 
b/devapidocs/org/apache/hadoop/hbase/HTableDescriptor.html
index bb8fd51..6831fb2 100644
--- a/devapidocs/org/apache/hadoop/hbase/HTableDescriptor.html
+++ b/devapidocs/org/apache/hadoop/hbase/HTableDescriptor.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":42,"i1":42,"i2":42,"i3":42,"i4":42,"i5":42,"i6":42,"i7":42,"i8":42,"i9":42,"i10":42,"i11":42,"i12":42,"i13":42,"i14":42,"i15":42,"i16":42,"i17":42,"i18":42,"i19":42,"i20":42,"i21":42,"i22":42,"i23":42,"i24":42,"i25":42,"i26":42,"i27":42,"i28":42,"i29":42,"i30":42,"i31":42,"i32":42,"i33":42,"i34":42,"i35":42,"i36":42,"i37":42,"i38":42,"i39":41,"i40":42,"i41":42,"i42":42,"i43":42,"i44":42,"i45":42,"i46":42,"i47":42,"i48":42,"i49":42,"i50":42,"i51":42,"i52":42,"i53":42,"i54":42,"i55":42,"i56":42,"i57":42,"i58":42,"i59":42,"i60":42,"i61":42,"i62":42,"i63":42,"i64":42,"i65":42,"i66":42};
+var methods = 
{"i0":42,"i1":42,"i2":42,"i3":42,"i4":42,"i5":42,"i6":42,"i7":42,"i8":42,"i9":42,"i10":42,"i11":42,"i12":42,"i13":42,"i14":42,"i15":42,"i16":42,"i17":42,"i18":42,"i19":42,"i20":42,"i21":42,"i22":42,"i23":42,"i24":42,"i25":42,"i26":42,"i27":42,"i28":42,"i29":42,"i30":42,"i31":42,"i32":42,"i33":42,"i34":42,"i35":42,"i36":42,"i37":42,"i38":42,"i39":42,"i40":42,"i41":42,"i42":42,"i43":42,"i44":41,"i45":42,"i46":42,"i47":42,"i48":42,"i49":42,"i50":42,"i51":42,"i52":42,"i53":42,"i54":42,"i55":42,"i56":42,"i57":42,"i58":42,"i59":42,"i60":42,"i61":42,"i62":42,"i63":42,"i64":42,"i65":42,"i66":42,"i67":42,"i68":42,"i69":42,"i70":42,"i71":42,"i72":42};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete 
Methods"],32:["t6","Deprecated Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -123,7 +123,7 @@ var activeTableTab = "activeTableTab";
 
 http://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true;
 title="class or interface in java.lang">@Deprecated
  @InterfaceAudience.Public
-public class HTableDescriptor
+public class HTableDescriptor
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements TableDescriptor, http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableHTableDescriptor
 HTableDescriptor contains the details about an HBase table  
such as the descriptors of
@@ -191,7 +191,7 @@ implements 
-private TableDescriptorBuilder.ModifyableTableDescriptor
+protected TableDescriptorBuilder.ModifyableTableDescriptor
 delegatee
 Deprecated.
 
@@ -306,6 +306,13 @@ implements 
+
+
+
+Fields inherited from interfaceorg.apache.hadoop.hbase.client.TableDescriptor
+COMPARATOR
+
 
 
 
@@ -328,33 +335,31 @@ implements 
+protected 
+HTableDescriptor(HTableDescriptordesc,
+booleandeepClone)
+Deprecated.
+
+
+
 
 HTableDescriptor(TableDescriptordesc)
 Deprecated.
 
 
-
+
 protected 
 HTableDescriptor(TableDescriptorBuilder.ModifyableTableDescriptordelegatee)
 Deprecated.
 
 
-
+
 
 HTableDescriptor(TableNamename)
 Deprecated.
 Construct a table descriptor specifying a TableName 
object
 
 
-
-private 
-HTableDescriptor(TableNamename,
-http://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
 title="class or interface in java.util">CollectionHColumnDescriptorfamilies,
-http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">MapBytes,Bytesvalues,
-http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">Stringconfiguration)
-Deprecated.
-
-
 
 
 HTableDescriptor(TableNamename,
@@ -428,69 +433,89 @@ implements HColumnDescriptor[]
 getColumnFamilies()
 Deprecated.
-Returns an array all the HColumnDescriptor of the column 
families
- of the table.
 
 
 
+ColumnFamilyDescriptor
+getColumnFamily(byte[]name)
+Deprecated.
+Returns the ColumnDescriptor for a specific column family 
with name as
+ specified by the parameter column.
+
+
+
 int
 getColumnFamilyCount()
 Deprecated.
 Returns the count of the column families of the table.
 
 
-
+
+http://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true;
 title="class or interface in java.util">Setbyte[]
+getColumnFamilyNames()
+Deprecated.
+Returns all the column family names of the current 
table.

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-06-07 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b44796ef/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html
--
diff --git a/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html 
b/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html
index b7c90be..f26f8aa 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/client/Scan.html
@@ -193,992 +193,1035 @@
 185   * Control whether to use pread at 
server side.
 186   */
 187  private ReadType readType = 
ReadType.DEFAULT;
-188  /**
-189   * Create a Scan operation across all 
rows.
-190   */
-191  public Scan() {}
-192
-193  /**
-194   * @deprecated use {@code new 
Scan().withStartRow(startRow).setFilter(filter)} instead.
-195   */
-196  @Deprecated
-197  public Scan(byte[] startRow, Filter 
filter) {
-198this(startRow);
-199this.filter = filter;
-200  }
-201
-202  /**
-203   * Create a Scan operation starting at 
the specified row.
-204   * p
-205   * If the specified row does not exist, 
the Scanner will start from the next closest row after the
-206   * specified row.
-207   * @param startRow row to start scanner 
at or after
-208   * @deprecated use {@code new 
Scan().withStartRow(startRow)} instead.
-209   */
-210  @Deprecated
-211  public Scan(byte[] startRow) {
-212setStartRow(startRow);
-213  }
-214
-215  /**
-216   * Create a Scan operation for the 
range of rows specified.
-217   * @param startRow row to start scanner 
at or after (inclusive)
-218   * @param stopRow row to stop scanner 
before (exclusive)
-219   * @deprecated use {@code new 
Scan().withStartRow(startRow).withStopRow(stopRow)} instead.
-220   */
-221  @Deprecated
-222  public Scan(byte[] startRow, byte[] 
stopRow) {
-223setStartRow(startRow);
-224setStopRow(stopRow);
-225  }
-226
-227  /**
-228   * Creates a new instance of this class 
while copying all values.
-229   *
-230   * @param scan  The scan instance to 
copy from.
-231   * @throws IOException When copying the 
values fails.
-232   */
-233  public Scan(Scan scan) throws 
IOException {
-234startRow = scan.getStartRow();
-235includeStartRow = 
scan.includeStartRow();
-236stopRow  = scan.getStopRow();
-237includeStopRow = 
scan.includeStopRow();
-238maxVersions = 
scan.getMaxVersions();
-239batch = scan.getBatch();
-240storeLimit = 
scan.getMaxResultsPerColumnFamily();
-241storeOffset = 
scan.getRowOffsetPerColumnFamily();
-242caching = scan.getCaching();
-243maxResultSize = 
scan.getMaxResultSize();
-244cacheBlocks = 
scan.getCacheBlocks();
-245filter = scan.getFilter(); // 
clone?
-246loadColumnFamiliesOnDemand = 
scan.getLoadColumnFamiliesOnDemandValue();
-247consistency = 
scan.getConsistency();
-248
this.setIsolationLevel(scan.getIsolationLevel());
-249reversed = scan.isReversed();
-250asyncPrefetch = 
scan.isAsyncPrefetch();
-251small = scan.isSmall();
-252allowPartialResults = 
scan.getAllowPartialResults();
-253tr = scan.getTimeRange(); // 
TimeRange is immutable
-254Mapbyte[], 
NavigableSetbyte[] fams = scan.getFamilyMap();
-255for 
(Map.Entrybyte[],NavigableSetbyte[] entry : fams.entrySet()) 
{
-256  byte [] fam = entry.getKey();
-257  NavigableSetbyte[] cols = 
entry.getValue();
-258  if (cols != null  
cols.size()  0) {
-259for (byte[] col : cols) {
-260  addColumn(fam, col);
-261}
-262  } else {
-263addFamily(fam);
-264  }
-265}
-266for (Map.EntryString, byte[] 
attr : scan.getAttributesMap().entrySet()) {
-267  setAttribute(attr.getKey(), 
attr.getValue());
+188
+189  private boolean needCursorResult = 
false;
+190
+191  /**
+192   * Create a Scan operation across all 
rows.
+193   */
+194  public Scan() {}
+195
+196  /**
+197   * @deprecated use {@code new 
Scan().withStartRow(startRow).setFilter(filter)} instead.
+198   */
+199  @Deprecated
+200  public Scan(byte[] startRow, Filter 
filter) {
+201this(startRow);
+202this.filter = filter;
+203  }
+204
+205  /**
+206   * Create a Scan operation starting at 
the specified row.
+207   * p
+208   * If the specified row does not exist, 
the Scanner will start from the next closest row after the
+209   * specified row.
+210   * @param startRow row to start scanner 
at or after
+211   * @deprecated use {@code new 
Scan().withStartRow(startRow)} instead.
+212   */
+213  @Deprecated
+214  public Scan(byte[] startRow) {
+215setStartRow(startRow);
+216  }
+217
+218  /**
+219   * Create a Scan operation for the 
range of rows specified.
+220   * @param startRow row to start scanner 
at or after (inclusive)
+221   * @param stopRow row to stop scanner 
before (exclusive)
+222   * @deprecated use {@code new 
Scan().withStartRow(startRow).withStopRow(stopRow)} instead.
+223   */
+224  @Deprecated
+225  public Scan(byte[] startRow, byte[] 
stopRow) 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-06-06 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6ad4f21a/devapidocs/org/apache/hadoop/hbase/master/procedure/DisableTableProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/DisableTableProcedure.html
 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/DisableTableProcedure.html
index d98b54f..1c410ff 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/DisableTableProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/DisableTableProcedure.html
@@ -389,7 +389,7 @@ extends Procedure
-addStackIndex,
 afterReplay,
 beforeReplay,
 compareTo,
 completionCleanup,
 doAcquireLock,
 doExecute,
 d
 oReleaseLock, doRollback,
 elapsedTime,
 getChildrenLatch,
 getException,
 getLastUpdate,
 getNonceKey,
 getOwner,
 getParentProcId,
 getProcId, getProcIdHashCode,
 getResult,
 getRootProcedureId,
 getRootProcId,
 getStackIndexes,
 getState,
 getSubmittedTime,
 getTimeout,
 <
 a 
href="../../../../../../org/apache/hadoop/hbase/procedure2/Procedure.html#getTimeoutTimestamp--">getTimeoutTimestamp,
 hasChildren,
 hasException,
 hasLock,
 hasOwner,
 hasParent,
 hasTimeout,
 haveSameParent,
 holdLock,
 incChildrenLatch,
 isFailed,
 isFinished,
 isInitializing,
 isRunnable,
 isSuccess,
 isWaiting,
 removeStackIndex,
 setAbortFailure,
 setChildrenLatch,
 setFailure,
 setFailure,
 setLastUpdate,
 setNonceKey,
 setOwner,
 setOwner,
 setParentProcId,
 setProcId,
 setResult,
 setRootProcId,
 setStackIndexes,
 setState,
 setSubmittedTime,
 setTimeout,
 setTimeoutFailure,
 shouldWaitClientAck,
 toString,
 toStringClass,
 toStringDetails,
 toStringSimpleSB,
 updateMetricsOnFinish,
 updateMetricsOnSubmit,
 updateTimestamp,
 wasExecuted
+addStackIndex,
 afterReplay,
 beforeReplay,
 compareTo,
 completionCleanup,
 doAcquireLock,
 doExecute,
 d
 oReleaseLock, doRollback,
 elapsedTime,
 getChildrenLatch,
 getException,
 getLastUpdate,
 getNonceKey,
 getOwner,
 getParentProcId,
 getProcedureMetrics, getProcId,
 getProcIdHashCode,
 getResult,
 getRootProcedureId,
 getRootProcId,
 getStackIndexes,
 getState,
 getSubmittedTime, getTimeout,
 getTimeoutTimestamp,
 hasChildren,
 hasException,
 hasLock,
 hasOwner,
 hasParent,
 hasTimeout,
 haveSameParent,
 holdLock,
 incChildrenLatch,
 isFailed,
 isFinished,
 isInitializing,
 isRunnable,
 isSuccess,
 isWaiting,
 removeStackIndex,
 setAbortFailure,
 setChildrenLatch,
 setFailure,
 setFailure,
 setLastUpdate,
 setNonceKey,
 setOwner, 
setOwner,
 setParentProcId,
 setProcId,
 setResult,
 setRootProcId,
 setStackIndexes,
 setState,
 setSubmittedTime,
 setTimeout,
 setTimeoutFailure,
 shouldWaitClientAck,
 toString,
 toStringClass,
 toStringDetails,
 toStringSimpleSB,
 updateMetricsOnFinish,
 updateMetricsOnSubmit,
 updateTimestamp,
 wasExecuted
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6ad4f21a/devapidocs/org/apache/hadoop/hbase/master/procedure/DispatchMergingRegionsProcedure.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/DispatchMergingRegionsProcedure.html
 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/DispatchMergingRegionsProcedure.html
index 5708c32..6734a0e 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/DispatchMergingRegionsProcedure.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/DispatchMergingRegionsProcedure.html
@@ -436,7 +436,7 @@ extends Procedure
-addStackIndex,
 afterReplay,
 beforeReplay,
 compareTo,
 completionCleanup,
 doAcquireLock,
 doExecute,
 d
 oReleaseLock, doRollback,
 elapsedTime,
 getChildrenLatch,
 getException,
 getLastUpdate,
 getNonceKey,
 getOwner,
 getParentProcId,
 getProcId, getProcIdHashCode,
 getResult,
 getRootProcedureId,
 getRootProcId,
 getStackIndexes,
 getState,
 getSubmittedTime,
 getTimeout,
 <
 a 
href="../../../../../../org/apache/hadoop/hbase/procedure2/Procedure.html#getTimeoutTimestamp--">getTimeoutTimestamp,
 hasChildren,
 hasException,
 hasLock,
 hasOwner,
 hasParent,
 hasTimeout,
 haveSameParent,
 holdLock,
 incChildrenLatch,
 isFailed,
 isFinished,
 isInitializing,
 isRunnable,
 isSuccess,
 isWaiting,
 removeStackIndex,
 setAbortFailure,
 setChildrenLatch,
 setFailure,
 setFailure,
 setLastUpdate,
 setNonceKey,
 setOwner,
 setOwner,
 setParentProcId,
 setProcId,
 setResult,
 setRootProcId,
 setStackIndexes,
 setState,
 setSubmittedTime,
 setTimeout,
 setTimeoutFailure,
 shouldWaitClientAck,
 toString,
 toStringClass,
 toStringDetails,
 toStringSimpleSB,
 updateMetricsOnFinish,
 updateMetricsOnSubmit,
 updateTimestamp,
 wasExecuted
+addStackIndex,
 afterReplay,
 beforeReplay,
 compareTo,
 completionCleanup,
 doAcquireLock,
 doExecute,
 d
 oReleaseLock, doRollback,
 elapsedTime,
 getChildrenLatch,
 getException,
 getLastUpdate,
 getNonceKey,
 getOwner,
 getParentProcId,
 getProcedureMetrics, getProcId,
 getProcIdHashCode,
 getResult,
 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-06-03 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9d35424/devapidocs/src-html/org/apache/hadoop/hbase/io/MetricsIOSource.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/io/MetricsIOSource.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/io/MetricsIOSource.html
new file mode 100644
index 000..b074547
--- /dev/null
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/io/MetricsIOSource.html
@@ -0,0 +1,152 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+Source code
+
+
+
+
+001/**
+002 * Licensed to the Apache Software 
Foundation (ASF) under one
+003 * or more contributor license 
agreements.  See the NOTICE file
+004 * distributed with this work for 
additional information
+005 * regarding copyright ownership.  The 
ASF licenses this file
+006 * to you under the Apache License, 
Version 2.0 (the
+007 * "License"); you may not use this file 
except in compliance
+008 * with the License.  You may obtain a 
copy of the License at
+009 *
+010 * 
http://www.apache.org/licenses/LICENSE-2.0
+011 *
+012 * Unless required by applicable law or 
agreed to in writing, software
+013 * distributed under the License is 
distributed on an "AS IS" BASIS,
+014 * WITHOUT WARRANTIES OR CONDITIONS OF 
ANY KIND, either express or implied.
+015 * See the License for the specific 
language governing permissions and
+016 * limitations under the License.
+017 */
+018
+019package org.apache.hadoop.hbase.io;
+020
+021import 
org.apache.hadoop.hbase.metrics.BaseSource;
+022
+023public interface MetricsIOSource extends 
BaseSource {
+024
+025  /**
+026   * The name of the metrics
+027   */
+028  String METRICS_NAME = "IO";
+029
+030  /**
+031   * The name of the metrics context that 
metrics will be under.
+032   */
+033  String METRICS_CONTEXT = 
"regionserver";
+034
+035  /**
+036   * Description
+037   */
+038  String METRICS_DESCRIPTION = "Metrics 
about FileSystem IO";
+039
+040  /**
+041   * The name of the metrics context that 
metrics will be under in jmx
+042   */
+043  String METRICS_JMX_CONTEXT = 
"RegionServer,sub=" + METRICS_NAME;
+044
+045
+046  String FS_READ_TIME_HISTO_KEY = 
"fsReadTime";
+047  String FS_PREAD_TIME_HISTO_KEY = 
"fsPReadTime";
+048  String FS_WRITE_HISTO_KEY = 
"fsWriteTime";
+049
+050  String CHECKSUM_FAILURES_KEY = 
"fsChecksumFailureCount";
+051
+052  String FS_READ_TIME_HISTO_DESC
+053= "Latency of HFile's sequential 
reads on this region server in milliseconds";
+054  String FS_PREAD_TIME_HISTO_DESC
+055= "Latency of HFile's positional 
reads on this region server in milliseconds";
+056  String FS_WRITE_TIME_HISTO_DESC
+057= "Latency of HFile's writes on this 
region server in milliseconds";
+058
+059  String CHECKSUM_FAILURES_DESC = "Number 
of checksum failures for the HBase HFile checksums at the"
+060  + " HBase level (separate from HDFS 
checksums)";
+061
+062
+063  /**
+064   * Update the fs sequential read time 
histogram
+065   * @param t time it took, in 
milliseconds
+066   */
+067  void updateFsReadTime(long t);
+068
+069  /**
+070   * Update the fs positional read time 
histogram
+071   * @param t time it took, in 
milliseconds
+072   */
+073  void updateFsPReadTime(long t);
+074
+075  /**
+076   * Update the fs write time histogram
+077   * @param t time it took, in 
milliseconds
+078   */
+079  void updateFsWriteTime(long t);
+080}
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9d35424/devapidocs/src-html/org/apache/hadoop/hbase/io/MetricsIOSourceImpl.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/io/MetricsIOSourceImpl.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/io/MetricsIOSourceImpl.html
new file mode 100644
index 000..11c20e6
--- /dev/null
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/io/MetricsIOSourceImpl.html
@@ -0,0 +1,156 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+Source code
+
+
+
+
+001/**
+002 * Licensed to the Apache Software 
Foundation (ASF) under one
+003 * or more contributor license 
agreements.  See the NOTICE file
+004 * distributed with this work for 
additional information
+005 * regarding copyright ownership.  The 
ASF licenses this file
+006 * to you under the Apache License, 
Version 2.0 (the
+007 * "License"); you may not use this file 
except in compliance
+008 * with the License.  You may obtain a 
copy of the License at
+009 *
+010 * 
http://www.apache.org/licenses/LICENSE-2.0
+011 *
+012 * Unless required by applicable law or 
agreed to in writing, software
+013 * distributed under the License is 
distributed on an "AS IS" BASIS,
+014 * WITHOUT WARRANTIES OR CONDITIONS OF 
ANY KIND, either express or implied.
+015 * See the License for the specific 
language governing permissions and
+016 * 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-06-01 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6dd31117/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html 
b/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html
index 025ba8d..25b805a 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/ShortCircuitMasterConnection.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -243,301 +243,311 @@ implements 
+org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsResponse
+dispatchMergingRegions(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
+  
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DispatchMergingRegionsRequestrequest)
+
+
 org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse
 drainRegionServers(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
   
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequestrequest)
 
-
+
 org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse
 enableCatalogJanitor(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequestrequest)
 
-
+
 org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse
 enableReplicationPeer(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
  
org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequestrequest)
 
-
+
 org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse
 enableTable(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,

org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequestrequest)
 
-
+
 org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse
 execMasterService(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
  
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequestrequest)
 
-
+
 org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse
 execProcedure(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
  
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequestrequest)
 
-
+
 org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse
 execProcedureWithRet(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequestrequest)
 
-
+
 org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse
 getClusterStatus(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllercontroller,
 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequestrequest)
 
-
+
 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dab57116/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.html 
b/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.html
index a6ce1bc..c98e496 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/ConnectionImplementation.html
@@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-class ConnectionImplementation
+class ConnectionImplementation
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements ClusterConnection, http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true;
 title="class or interface in java.io">Closeable
 Main implementation of Connection 
and ClusterConnection interfaces.
@@ -857,7 +857,7 @@ implements 
 
 RETRIES_BY_SERVER_KEY
-public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String RETRIES_BY_SERVER_KEY
+public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String RETRIES_BY_SERVER_KEY
 
 See Also:
 Constant
 Field Values
@@ -870,7 +870,7 @@ implements 
 
 LOG
-private static finalorg.apache.commons.logging.Log LOG
+private static finalorg.apache.commons.logging.Log LOG
 
 
 
@@ -879,7 +879,7 @@ implements 
 
 RESOLVE_HOSTNAME_ON_FAIL_KEY
-private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String RESOLVE_HOSTNAME_ON_FAIL_KEY
+private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String RESOLVE_HOSTNAME_ON_FAIL_KEY
 
 See Also:
 Constant
 Field Values
@@ -892,7 +892,7 @@ implements 
 
 hostnamesCanChange
-private finalboolean hostnamesCanChange
+private finalboolean hostnamesCanChange
 
 
 
@@ -901,7 +901,7 @@ implements 
 
 pause
-private finallong pause
+private finallong pause
 
 
 
@@ -910,7 +910,7 @@ implements 
 
 pauseForCQTBE
-private finallong pauseForCQTBE
+private finallong pauseForCQTBE
 
 
 
@@ -919,7 +919,7 @@ implements 
 
 useMetaReplicas
-privateboolean useMetaReplicas
+privateboolean useMetaReplicas
 
 
 
@@ -928,7 +928,7 @@ implements 
 
 metaReplicaCallTimeoutScanInMicroSecond
-private finalint metaReplicaCallTimeoutScanInMicroSecond
+private finalint metaReplicaCallTimeoutScanInMicroSecond
 
 
 
@@ -937,7 +937,7 @@ implements 
 
 numTries
-private finalint numTries
+private finalint numTries
 
 
 
@@ -946,7 +946,7 @@ implements 
 
 rpcTimeout
-finalint rpcTimeout
+finalint rpcTimeout
 
 
 
@@ -955,7 +955,7 @@ implements 
 
 nonceGenerator
-private static volatileNonceGenerator nonceGenerator
+private static volatileNonceGenerator nonceGenerator
 Global nonceGenerator shared per client.Currently there's 
no reason to limit its scope.
  Once it's set under nonceGeneratorCreateLock, it is never unset or 
changed.
 
@@ -966,7 +966,7 @@ implements 
 
 nonceGeneratorCreateLock
-private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object nonceGeneratorCreateLock
+private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object nonceGeneratorCreateLock
 The nonce generator lock. Only taken when creating 
Connection, which gets a private copy.
 
 
@@ -976,7 +976,7 @@ implements 
 
 asyncProcess
-private finalAsyncProcess asyncProcess
+private finalAsyncProcess asyncProcess
 
 
 
@@ -985,7 +985,7 @@ implements 
 
 stats
-private finalServerStatisticTracker stats
+private finalServerStatisticTracker stats
 
 
 
@@ -994,7 +994,7 @@ implements 
 
 closed
-private volatileboolean closed
+private volatileboolean closed
 
 
 
@@ -1003,7 +1003,7 @@ implements 
 
 aborted
-private volatileboolean aborted
+private volatileboolean aborted
 
 
 
@@ -1012,7 +1012,7 @@ implements 
 
 clusterStatusListener
-ClusterStatusListener clusterStatusListener
+ClusterStatusListener clusterStatusListener
 
 
 
@@ -1021,7 +1021,7 @@ implements 
 
 metaRegionLock
-private finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object metaRegionLock
+private finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object metaRegionLock
 
 
 
@@ -1030,7 +1030,7 @@ implements 
 
 masterAndZKLock
-private 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-19 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c635e71b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
index be7f8e5..37574d7 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/ConnectionImplementation.ServerErrorTracker.ServerErrors.html
@@ -147,1885 +147,1897 @@
 139  private final boolean 
hostnamesCanChange;
 140  private final long pause;
 141  private final long pauseForCQTBE;// 
pause for CallQueueTooBigException, if specified
-142  private final boolean 
useMetaReplicas;
-143  private final int numTries;
-144  final int rpcTimeout;
-145
-146  /**
-147   * Global nonceGenerator shared per 
client.Currently there's no reason to limit its scope.
-148   * Once it's set under 
nonceGeneratorCreateLock, it is never unset or changed.
-149   */
-150  private static volatile NonceGenerator 
nonceGenerator = null;
-151  /** The nonce generator lock. Only 
taken when creating Connection, which gets a private copy. */
-152  private static final Object 
nonceGeneratorCreateLock = new Object();
-153
-154  private final AsyncProcess 
asyncProcess;
-155  // single tracker per connection
-156  private final ServerStatisticTracker 
stats;
-157
-158  private volatile boolean closed;
-159  private volatile boolean aborted;
-160
-161  // package protected for the tests
-162  ClusterStatusListener 
clusterStatusListener;
-163
-164  private final Object metaRegionLock = 
new Object();
-165
-166  // We have a single lock for master 
 zk to prevent deadlocks. Having
-167  //  one lock for ZK and one lock for 
master is not possible:
-168  //  When creating a connection to 
master, we need a connection to ZK to get
-169  //  its address. But another thread 
could have taken the ZK lock, and could
-170  //  be waiting for the master lock 
= deadlock.
-171  private final Object masterAndZKLock = 
new Object();
-172
-173  // thread executor shared by all Table 
instances created
-174  // by this connection
-175  private volatile ExecutorService 
batchPool = null;
-176  // meta thread executor shared by all 
Table instances created
-177  // by this connection
-178  private volatile ExecutorService 
metaLookupPool = null;
-179  private volatile boolean cleanupPool = 
false;
-180
-181  private final Configuration conf;
-182
-183  // cache the configuration value for 
tables so that we can avoid calling
-184  // the expensive Configuration to fetch 
the value multiple times.
-185  private final ConnectionConfiguration 
connectionConfig;
-186
-187  // Client rpc instance.
-188  private final RpcClient rpcClient;
-189
-190  private final MetaCache metaCache;
-191  private final MetricsConnection 
metrics;
-192
-193  protected User user;
-194
-195  private final RpcRetryingCallerFactory 
rpcCallerFactory;
-196
-197  private final RpcControllerFactory 
rpcControllerFactory;
-198
-199  private final RetryingCallerInterceptor 
interceptor;
-200
-201  /**
-202   * Cluster registry of basic info such 
as clusterid and meta region location.
-203   */
-204  Registry registry;
-205
-206  private final ClientBackoffPolicy 
backoffPolicy;
-207
-208  /**
-209   * Allow setting an alternate 
BufferedMutator implementation via
-210   * config. If null, use default.
-211   */
-212  private final String 
alternateBufferedMutatorClassName;
-213
-214  /**
-215   * constructor
-216   * @param conf Configuration object
-217   */
-218  ConnectionImplementation(Configuration 
conf,
-219   
ExecutorService pool, User user) throws IOException {
-220this.conf = conf;
-221this.user = user;
-222this.batchPool = pool;
-223this.connectionConfig = new 
ConnectionConfiguration(conf);
-224this.closed = false;
-225this.pause = 
conf.getLong(HConstants.HBASE_CLIENT_PAUSE,
-226
HConstants.DEFAULT_HBASE_CLIENT_PAUSE);
-227long configuredPauseForCQTBE = 
conf.getLong(HConstants.HBASE_CLIENT_PAUSE_FOR_CQTBE, pause);
-228if (configuredPauseForCQTBE  
pause) {
-229  LOG.warn("The " + 
HConstants.HBASE_CLIENT_PAUSE_FOR_CQTBE + " setting: "
-230  + configuredPauseForCQTBE + " 
is smaller than " + HConstants.HBASE_CLIENT_PAUSE
-231  + ", will use " + pause + " 
instead.");
-232  this.pauseForCQTBE = pause;
-233} else {
-234  this.pauseForCQTBE = 
configuredPauseForCQTBE;
-235}
-236this.useMetaReplicas = 
conf.getBoolean(HConstants.USE_META_REPLICAS,
-237  
HConstants.DEFAULT_USE_META_REPLICAS);
-238// how many times to try, one more 
than max *retry* 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-18 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8b4cf63f/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
index f81384b..f1c87a7 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
@@ -1192,6 +1192,14 @@
 
 
 private void
+CloneSnapshotProcedure.restoreSnapshotAcl(MasterProcedureEnvenv)
+
+
+private void
+RestoreSnapshotProcedure.restoreSnapshotAcl(MasterProcedureEnvenv)
+
+
+private void
 ModifyColumnFamilyProcedure.restoreTableDescriptor(MasterProcedureEnvenv)
 Restore back to the old descriptor
 
@@ -1637,59 +1645,65 @@
 http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListHRegionInforegions)
 
 
-CloneSnapshotProcedure(MasterProcedureEnvenv,
+CloneSnapshotProcedure(MasterProcedureEnvenv,
   HTableDescriptorhTableDescriptor,
-  
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.SnapshotDescriptionsnapshot)
+  
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionsnapshot)
+
+
+CloneSnapshotProcedure(MasterProcedureEnvenv,
+  HTableDescriptorhTableDescriptor,
+  
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionsnapshot,
+  booleanrestoreAcl)
 Constructor
 
 
-
+
 CreateNamespaceProcedure(MasterProcedureEnvenv,
 NamespaceDescriptornsDescriptor)
 
-
+
 CreateTableProcedure(MasterProcedureEnvenv,
 HTableDescriptorhTableDescriptor,
 HRegionInfo[]newRegions)
 
-
+
 CreateTableProcedure(MasterProcedureEnvenv,
 HTableDescriptorhTableDescriptor,
 HRegionInfo[]newRegions,
 ProcedurePrepareLatchsyncLatch)
 
-
+
 DeleteColumnFamilyProcedure(MasterProcedureEnvenv,
TableNametableName,
byte[]familyName)
 
-
+
 DeleteColumnFamilyProcedure(MasterProcedureEnvenv,
TableNametableName,
byte[]familyName,
ProcedurePrepareLatchlatch)
 
-
+
 DeleteNamespaceProcedure(MasterProcedureEnvenv,
 http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">StringnamespaceName)
 
-
+
 DeleteTableProcedure(MasterProcedureEnvenv,
 TableNametableName)
 
-
+
 DeleteTableProcedure(MasterProcedureEnvenv,
 TableNametableName,
 ProcedurePrepareLatchsyncLatch)
 
-
+
 DisableTableProcedure(MasterProcedureEnvenv,
  TableNametableName,
  booleanskipTableStateCheck)
 Constructor
 
 
-
+
 DisableTableProcedure(MasterProcedureEnvenv,
  TableNametableName,
  booleanskipTableStateCheck,
@@ -1697,14 +1711,14 @@
 Constructor
 
 
-
+
 EnableTableProcedure(MasterProcedureEnvenv,
 TableNametableName,
 booleanskipTableStateCheck)
 Constructor
 
 
-
+
 EnableTableProcedure(MasterProcedureEnvenv,
 TableNametableName,
 booleanskipTableStateCheck,
@@ -1712,39 +1726,45 @@
 Constructor
 
 
-
+
 MergeTableRegionsProcedure(MasterProcedureEnvenv,
   HRegionInfo[]regionsToMerge,
   booleanforcible)
 
-
+
 ModifyColumnFamilyProcedure(MasterProcedureEnvenv,
TableNametableName,
HColumnDescriptorcfDescriptor)
 
-
+
 ModifyColumnFamilyProcedure(MasterProcedureEnvenv,
TableNametableName,
HColumnDescriptorcfDescriptor,
ProcedurePrepareLatchlatch)
 
-
+
 ModifyNamespaceProcedure(MasterProcedureEnvenv,
 NamespaceDescriptornewNsDescriptor)
 
-
+
 ModifyTableProcedure(MasterProcedureEnvenv,
 HTableDescriptorhtd)
 
-
+
 ModifyTableProcedure(MasterProcedureEnvenv,
 HTableDescriptorhtd,
 ProcedurePrepareLatchlatch)
 
+
+RestoreSnapshotProcedure(MasterProcedureEnvenv,
+HTableDescriptorhTableDescriptor,
+
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionsnapshot)
+
 
-RestoreSnapshotProcedure(MasterProcedureEnvenv,

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-15 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f55ebeaa/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html
index 109b5f3..e484176 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.MessageDecoder.html
@@ -23,484 +23,310 @@
 015 * See the License for the specific 
language governing permissions and
 016 * limitations under the License.
 017 */
-018
-019package org.apache.hadoop.hbase.ipc;
-020
-021import 
io.netty.bootstrap.ServerBootstrap;
-022import io.netty.buffer.ByteBuf;
-023import 
io.netty.buffer.PooledByteBufAllocator;
-024import io.netty.buffer.Unpooled;
-025import io.netty.channel.Channel;
-026import io.netty.channel.ChannelFuture;
-027import 
io.netty.channel.ChannelFutureListener;
-028import 
io.netty.channel.ChannelHandlerContext;
-029import 
io.netty.channel.ChannelInboundHandlerAdapter;
-030import 
io.netty.channel.ChannelInitializer;
-031import io.netty.channel.ChannelOption;
-032import 
io.netty.channel.ChannelOutboundHandlerAdapter;
-033import 
io.netty.channel.ChannelPipeline;
-034import io.netty.channel.ChannelPromise;
-035import io.netty.channel.EventLoopGroup;
-036import 
io.netty.channel.epoll.EpollEventLoopGroup;
-037import 
io.netty.channel.epoll.EpollServerSocketChannel;
-038import 
io.netty.channel.group.ChannelGroup;
-039import 
io.netty.channel.group.DefaultChannelGroup;
-040import 
io.netty.channel.nio.NioEventLoopGroup;
-041import 
io.netty.channel.socket.SocketChannel;
-042import 
io.netty.channel.socket.nio.NioServerSocketChannel;
-043import 
io.netty.handler.codec.ByteToMessageDecoder;
-044import 
io.netty.handler.codec.LengthFieldBasedFrameDecoder;
-045import 
io.netty.util.concurrent.GlobalEventExecutor;
-046
-047import java.io.IOException;
-048import java.io.InterruptedIOException;
-049import java.net.InetAddress;
-050import java.net.InetSocketAddress;
-051import java.nio.ByteBuffer;
-052import java.util.Arrays;
-053import java.util.List;
-054import 
java.util.concurrent.CountDownLatch;
-055
-056import org.apache.commons.logging.Log;
-057import 
org.apache.commons.logging.LogFactory;
-058import 
org.apache.hadoop.conf.Configuration;
-059import 
org.apache.hadoop.hbase.CellScanner;
-060import 
org.apache.hadoop.hbase.HConstants;
-061import org.apache.hadoop.hbase.Server;
-062import 
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
-063import 
org.apache.hadoop.hbase.nio.ByteBuff;
-064import 
org.apache.hadoop.hbase.nio.SingleByteBuff;
-065import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-066import 
org.apache.hadoop.hbase.security.AuthMethod;
-067import 
org.apache.hadoop.hbase.security.HBasePolicyProvider;
-068import 
org.apache.hadoop.hbase.security.SaslStatus;
-069import 
org.apache.hadoop.hbase.security.SaslUtil;
-070import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
-071import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
-072import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
-073import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
-074import 
org.apache.hadoop.hbase.util.Bytes;
-075import 
org.apache.hadoop.hbase.util.JVM;
-076import 
org.apache.hadoop.hbase.util.Pair;
-077import 
org.apache.hadoop.io.IntWritable;
-078import 
org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
-079import org.apache.htrace.TraceInfo;
+018package org.apache.hadoop.hbase.ipc;
+019
+020import 
io.netty.bootstrap.ServerBootstrap;
+021import io.netty.buffer.ByteBuf;
+022import 
io.netty.buffer.PooledByteBufAllocator;
+023import io.netty.buffer.Unpooled;
+024import io.netty.channel.Channel;
+025import io.netty.channel.ChannelFuture;
+026import 
io.netty.channel.ChannelFutureListener;
+027import 
io.netty.channel.ChannelHandlerContext;
+028import 
io.netty.channel.ChannelInboundHandlerAdapter;
+029import 
io.netty.channel.ChannelInitializer;
+030import io.netty.channel.ChannelOption;
+031import 
io.netty.channel.ChannelOutboundHandlerAdapter;
+032import 
io.netty.channel.ChannelPipeline;
+033import io.netty.channel.ChannelPromise;
+034import io.netty.channel.EventLoopGroup;
+035import 
io.netty.channel.epoll.EpollEventLoopGroup;
+036import 
io.netty.channel.epoll.EpollServerSocketChannel;
+037import 
io.netty.channel.group.ChannelGroup;
+038import 
io.netty.channel.group.DefaultChannelGroup;
+039import 
io.netty.channel.nio.NioEventLoopGroup;
+040import 
io.netty.channel.socket.SocketChannel;
+041import 
io.netty.channel.socket.nio.NioServerSocketChannel;
+042import 
io.netty.handler.codec.ByteToMessageDecoder;
+043import 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-13 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8e0a5167/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html
index cc40d21..2df0b04 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupCommands.ProgressCommand.html
@@ -61,8 +61,8 @@
 053import 
org.apache.hadoop.hbase.backup.BackupRestoreConstants;
 054import 
org.apache.hadoop.hbase.backup.BackupRestoreConstants.BackupCommand;
 055import 
org.apache.hadoop.hbase.backup.BackupType;
-056import 
org.apache.hadoop.hbase.backup.util.BackupUtils;
-057import 
org.apache.hadoop.hbase.backup.util.BackupSet;
+056import 
org.apache.hadoop.hbase.backup.util.BackupSet;
+057import 
org.apache.hadoop.hbase.backup.util.BackupUtils;
 058import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
 059import 
org.apache.hadoop.hbase.client.Connection;
 060import 
org.apache.hadoop.hbase.client.ConnectionFactory;
@@ -122,670 +122,703 @@
 114
 115  public static abstract class Command 
extends Configured {
 116CommandLine cmdline;
-117
+117Connection conn;
 118Command(Configuration conf) {
-119  super(conf);
-120}
-121
-122public void execute() throws 
IOException {
-123  if (cmdline.hasOption("h") || 
cmdline.hasOption("help")) {
-124printUsage();
-125throw new 
IOException(INCORRECT_USAGE);
-126  }
-127}
-128
-129protected abstract void 
printUsage();
-130  }
-131
-132  private BackupCommands() {
-133throw new 
AssertionError("Instantiating utility class...");
-134  }
-135
-136  public static Command 
createCommand(Configuration conf, BackupCommand type, CommandLine cmdline) {
-137Command cmd = null;
-138switch (type) {
-139case CREATE:
-140  cmd = new CreateCommand(conf, 
cmdline);
-141  break;
-142case DESCRIBE:
-143  cmd = new DescribeCommand(conf, 
cmdline);
-144  break;
-145case PROGRESS:
-146  cmd = new ProgressCommand(conf, 
cmdline);
-147  break;
-148case DELETE:
-149  cmd = new DeleteCommand(conf, 
cmdline);
-150  break;
-151case CANCEL:
-152  cmd = new CancelCommand(conf, 
cmdline);
-153  break;
-154case HISTORY:
-155  cmd = new HistoryCommand(conf, 
cmdline);
-156  break;
-157case SET:
-158  cmd = new BackupSetCommand(conf, 
cmdline);
-159  break;
-160case HELP:
-161default:
-162  cmd = new HelpCommand(conf, 
cmdline);
-163  break;
-164}
-165return cmd;
-166  }
-167
-168  static int numOfArgs(String[] args) {
-169if (args == null) return 0;
-170return args.length;
-171  }
-172
-173  public static class CreateCommand 
extends Command {
-174
-175CreateCommand(Configuration conf, 
CommandLine cmdline) {
-176  super(conf);
-177  this.cmdline = cmdline;
-178}
-179
-180@Override
-181public void execute() throws 
IOException {
-182  super.execute();
-183  if (cmdline == null || 
cmdline.getArgs() == null) {
-184printUsage();
-185throw new 
IOException(INCORRECT_USAGE);
-186  }
-187  String[] args = 
cmdline.getArgs();
-188  if (args.length !=3) {
-189printUsage();
-190throw new 
IOException(INCORRECT_USAGE);
-191  }
-192
-193  if 
(!BackupType.FULL.toString().equalsIgnoreCase(args[1])
-194   
!BackupType.INCREMENTAL.toString().equalsIgnoreCase(args[1])) {
-195System.out.println("ERROR: 
invalid backup type: " + args[1]);
-196printUsage();
-197throw new 
IOException(INCORRECT_USAGE);
-198  }
-199  if (!verifyPath(args[2])) {
-200System.out.println("ERROR: 
invalid backup destination: " + args[2]);
-201printUsage();
-202throw new 
IOException(INCORRECT_USAGE);
-203  }
-204
-205  String tables = null;
-206  Configuration conf = getConf() != 
null ? getConf() : HBaseConfiguration.create();
-207
-208  // Check if we have both: backup 
set and list of tables
-209  if (cmdline.hasOption(OPTION_TABLE) 
 cmdline.hasOption(OPTION_SET)) {
-210System.out.println("ERROR: You 
can specify either backup set or list"
-211+ " of tables, but not 
both");
-212printUsage();
-213throw new 
IOException(INCORRECT_USAGE);
-214  }
-215
-216  // Check backup set
-217  String setName = null;
-218  if (cmdline.hasOption(OPTION_SET)) 
{
-219setName = 
cmdline.getOptionValue(OPTION_SET);
-220tables = getTablesForSet(setName, 
conf);
-221
-222if (tables == null) {
-223  System.out.println("ERROR: 
Backup set '" + setName
-224

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-12 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/709b8fcc/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html
--
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html
index d77512a..de81692 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/Admin.html
@@ -1543,47 +1543,47 @@
 1535
 1536  /**
 1537   * Create a new table by cloning the 
snapshot content.
-1538   * @param snapshotName name of the 
snapshot to be cloned
-1539   * @param tableName name of the table 
where the snapshot will be restored
-1540   * @param restoreAcl true to clone acl 
into newly created table
+1538   *
+1539   * @param snapshotName name of the 
snapshot to be cloned
+1540   * @param tableName name of the table 
where the snapshot will be restored
 1541   * @throws IOException if a remote or 
network exception occurs
 1542   * @throws TableExistsException if 
table to be created already exists
 1543   * @throws RestoreSnapshotException if 
snapshot failed to be cloned
 1544   * @throws IllegalArgumentException if 
the specified table has not a valid name
 1545   */
-1546  void cloneSnapshot(final String 
snapshotName, final TableName tableName, final boolean restoreAcl)
+1546  void cloneSnapshot(final String 
snapshotName, final TableName tableName)
 1547  throws IOException, 
TableExistsException, RestoreSnapshotException;
 1548
 1549  /**
-1550   * Create a new table by cloning the 
snapshot content.
-1551   *
-1552   * @param snapshotName name of the 
snapshot to be cloned
-1553   * @param tableName name of the table 
where the snapshot will be restored
-1554   * @throws IOException if a remote or 
network exception occurs
-1555   * @throws TableExistsException if 
table to be created already exists
-1556   * @throws RestoreSnapshotException if 
snapshot failed to be cloned
-1557   * @throws IllegalArgumentException if 
the specified table has not a valid name
-1558   */
-1559  void cloneSnapshot(final String 
snapshotName, final TableName tableName)
-1560  throws IOException, 
TableExistsException, RestoreSnapshotException;
-1561
-1562  /**
-1563   * Create a new table by cloning the 
snapshot content, but does not block
-1564   * and wait for it be completely 
cloned.
-1565   * You can use Future.get(long, 
TimeUnit) to wait on the operation to complete.
-1566   * It may throw ExecutionException if 
there was an error while executing the operation
-1567   * or TimeoutException in case the 
wait timeout was not long enough to allow the
-1568   * operation to complete.
+1550   * Create a new table by cloning the 
snapshot content, but does not block
+1551   * and wait for it be completely 
cloned.
+1552   * You can use Future.get(long, 
TimeUnit) to wait on the operation to complete.
+1553   * It may throw ExecutionException if 
there was an error while executing the operation
+1554   * or TimeoutException in case the 
wait timeout was not long enough to allow the
+1555   * operation to complete.
+1556   *
+1557   * @param snapshotName name of the 
snapshot to be cloned
+1558   * @param tableName name of the table 
where the snapshot will be restored
+1559   * @throws IOException if a remote or 
network exception occurs
+1560   * @throws TableExistsException if 
table to be cloned already exists
+1561   * @return the result of the async 
clone snapshot. You can use Future.get(long, TimeUnit)
+1562   *to wait on the operation to 
complete.
+1563   */
+1564  FutureVoid 
cloneSnapshotAsync(final String snapshotName, final TableName tableName)
+1565  throws IOException, 
TableExistsException;
+1566
+1567  /**
+1568   * Execute a distributed procedure on 
a cluster.
 1569   *
-1570   * @param snapshotName name of the 
snapshot to be cloned
-1571   * @param tableName name of the table 
where the snapshot will be restored
-1572   * @throws IOException if a remote or 
network exception occurs
-1573   * @throws TableExistsException if 
table to be cloned already exists
-1574   * @return the result of the async 
clone snapshot. You can use Future.get(long, TimeUnit)
-1575   *to wait on the operation to 
complete.
+1570   * @param signature A distributed 
procedure is uniquely identified by its signature (default the
+1571   * root ZK node name of the 
procedure).
+1572   * @param instance The instance name 
of the procedure. For some procedures, this parameter is
+1573   * optional.
+1574   * @param props Property/Value pairs 
of properties passing to the procedure
+1575   * @throws IOException
 1576   */
-1577  FutureVoid 
cloneSnapshotAsync(final String snapshotName, final TableName tableName)
-1578  throws IOException, 
TableExistsException;
+1577  void execProcedure(String signature, 
String instance, MapString, String props)
+1578  throws IOException;
 1579
 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dd7176bf/devapidocs/org/apache/hadoop/hbase/security/access/TablePermission.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/security/access/TablePermission.html 
b/devapidocs/org/apache/hadoop/hbase/security/access/TablePermission.html
index f16a58b..0f910b7 100644
--- a/devapidocs/org/apache/hadoop/hbase/security/access/TablePermission.html
+++ b/devapidocs/org/apache/hadoop/hbase/security/access/TablePermission.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -374,14 +374,18 @@ extends readFields(http://docs.oracle.com/javase/8/docs/api/java/io/DataInput.html?is-external=true;
 title="class or interface in java.io">DataInputin)
 
 
+void
+setTableName(TableNametable)
+
+
 boolean
 tableFieldsEqual(TablePermissionother)
 
-
+
 http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String
 toString()
 
-
+
 void
 write(http://docs.oracle.com/javase/8/docs/api/java/io/DataOutput.html?is-external=true;
 title="class or interface in 
java.io">DataOutputout)
 
@@ -631,13 +635,22 @@ extends TableNamegetTableName()
 
 
+
+
+
+
+
+setTableName
+publicvoidsetTableName(TableNametable)
+
+
 
 
 
 
 
 hasFamily
-publicbooleanhasFamily()
+publicbooleanhasFamily()
 
 
 
@@ -646,7 +659,7 @@ extends 
 
 getFamily
-publicbyte[]getFamily()
+publicbyte[]getFamily()
 
 
 
@@ -655,7 +668,7 @@ extends 
 
 hasQualifier
-publicbooleanhasQualifier()
+publicbooleanhasQualifier()
 
 
 
@@ -664,7 +677,7 @@ extends 
 
 getQualifier
-publicbyte[]getQualifier()
+publicbyte[]getQualifier()
 
 
 
@@ -673,7 +686,7 @@ extends 
 
 hasNamespace
-publicbooleanhasNamespace()
+publicbooleanhasNamespace()
 
 
 
@@ -682,7 +695,7 @@ extends 
 
 getNamespace
-publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringgetNamespace()
+publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringgetNamespace()
 
 
 
@@ -691,7 +704,7 @@ extends 
 
 implies
-publicbooleanimplies(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringnamespace,
+publicbooleanimplies(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringnamespace,
Permission.Actionaction)
 Checks that a given table operation is authorized by this 
permission
  instance.
@@ -711,7 +724,7 @@ extends 
 
 implies
-publicbooleanimplies(TableNametable,
+publicbooleanimplies(TableNametable,
byte[]family,
byte[]qualifier,
Permission.Actionaction)
@@ -737,7 +750,7 @@ extends 
 
 implies
-publicbooleanimplies(TableNametable,
+publicbooleanimplies(TableNametable,
KeyValuekv,
Permission.Actionaction)
 Checks if this permission grants access to perform the 
given action on
@@ -759,7 +772,7 @@ extends 
 
 matchesFamily
-publicbooleanmatchesFamily(TableNametable,
+publicbooleanmatchesFamily(TableNametable,
  byte[]family,
  Permission.Actionaction)
 Returns true if this permission matches the 
given column
@@ -776,7 +789,7 @@ extends 
 
 matchesFamilyQualifier
-publicbooleanmatchesFamilyQualifier(TableNametable,
+publicbooleanmatchesFamilyQualifier(TableNametable,
   byte[]family,
   byte[]qualifier,
   Permission.Actionaction)
@@ -799,7 +812,7 @@ extends 
 
 tableFieldsEqual
-publicbooleantableFieldsEqual(TablePermissionother)
+publicbooleantableFieldsEqual(TablePermissionother)
 
 
 
@@ -808,7 +821,7 @@ extends 
 
 equals
-publicbooleanequals(http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Objectobj)
+publicbooleanequals(http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Objectobj)
 
 Overrides:
 equalsin
 classPermission
@@ -821,7 +834,7 @@ extends 
 
 hashCode
-publicinthashCode()
+publicinthashCode()
 
 Overrides:
 hashCodein
 classPermission
@@ 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ef4c5a9/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html
index 27e0dee..109b5f3 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcServer.ConnectionHeaderHandler.html
@@ -67,157 +67,157 @@
 059import 
org.apache.hadoop.hbase.CellScanner;
 060import 
org.apache.hadoop.hbase.HConstants;
 061import org.apache.hadoop.hbase.Server;
-062import 
org.apache.hadoop.hbase.classification.InterfaceStability;
-063import 
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
-064import 
org.apache.hadoop.hbase.nio.ByteBuff;
-065import 
org.apache.hadoop.hbase.nio.SingleByteBuff;
-066import 
org.apache.hadoop.hbase.security.AccessDeniedException;
-067import 
org.apache.hadoop.hbase.security.AuthMethod;
-068import 
org.apache.hadoop.hbase.security.HBasePolicyProvider;
-069import 
org.apache.hadoop.hbase.security.SaslStatus;
-070import 
org.apache.hadoop.hbase.security.SaslUtil;
-071import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
-072import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
-073import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
-074import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
-075import 
org.apache.hadoop.hbase.util.Bytes;
-076import 
org.apache.hadoop.hbase.util.JVM;
-077import 
org.apache.hadoop.hbase.util.Pair;
-078import 
org.apache.hadoop.io.IntWritable;
-079import 
org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
-080import org.apache.htrace.TraceInfo;
-081
-082/**
-083 * An RPC server with Netty4 
implementation.
-084 *
-085 */
-086public class NettyRpcServer extends 
RpcServer {
-087
-088  public static final Log LOG = 
LogFactory.getLog(NettyRpcServer.class);
-089
-090  protected final InetSocketAddress 
bindAddress;
-091
-092  private final CountDownLatch closed = 
new CountDownLatch(1);
-093  private final Channel serverChannel;
-094  private final ChannelGroup allChannels 
= new DefaultChannelGroup(GlobalEventExecutor.INSTANCE);;
-095
-096  public NettyRpcServer(final Server 
server, final String name,
-097  final 
ListBlockingServiceAndInterface services,
-098  final InetSocketAddress 
bindAddress, Configuration conf,
-099  RpcScheduler scheduler) throws 
IOException {
-100super(server, name, services, 
bindAddress, conf, scheduler);
-101this.bindAddress = bindAddress;
-102boolean useEpoll = useEpoll(conf);
-103int workerCount = 
conf.getInt("hbase.netty.rpc.server.worker.count",
-104
Runtime.getRuntime().availableProcessors() / 4);
-105EventLoopGroup bossGroup = null;
-106EventLoopGroup workerGroup = null;
-107if (useEpoll) {
-108  bossGroup = new 
EpollEventLoopGroup(1);
-109  workerGroup = new 
EpollEventLoopGroup(workerCount);
-110} else {
-111  bossGroup = new 
NioEventLoopGroup(1);
-112  workerGroup = new 
NioEventLoopGroup(workerCount);
-113}
-114ServerBootstrap bootstrap = new 
ServerBootstrap();
-115bootstrap.group(bossGroup, 
workerGroup);
-116if (useEpoll) {
-117  
bootstrap.channel(EpollServerSocketChannel.class);
-118} else {
-119  
bootstrap.channel(NioServerSocketChannel.class);
-120}
-121
bootstrap.childOption(ChannelOption.TCP_NODELAY, tcpNoDelay);
-122
bootstrap.childOption(ChannelOption.SO_KEEPALIVE, tcpKeepAlive);
-123
bootstrap.childOption(ChannelOption.ALLOCATOR,
-124
PooledByteBufAllocator.DEFAULT);
-125bootstrap.childHandler(new 
Initializer(maxRequestSize));
-126
-127try {
-128  serverChannel = 
bootstrap.bind(this.bindAddress).sync().channel();
-129  LOG.info("NettyRpcServer bind to 
address=" + serverChannel.localAddress()
-130  + ", 
hbase.netty.rpc.server.worker.count=" + workerCount
-131  + ", useEpoll=" + useEpoll);
-132  allChannels.add(serverChannel);
-133} catch (InterruptedException e) {
-134  throw new 
InterruptedIOException(e.getMessage());
-135}
-136initReconfigurable(conf);
-137this.scheduler.init(new 
RpcSchedulerContext(this));
-138  }
-139
-140  private static boolean 
useEpoll(Configuration conf) {
-141// Config to enable native 
transport.
-142boolean epollEnabled = 
conf.getBoolean("hbase.rpc.server.nativetransport",
-143true);
-144// Use the faster native epoll 
transport mechanism on linux if enabled
-145return epollEnabled  
JVM.isLinux()  JVM.isAmd64();
-146  }
-147
-148  @Override
-149  public synchronized void start() {
-150if 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-05-03 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/31df4674/devapidocs/org/apache/hadoop/hbase/regionserver/StoreScanner.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/StoreScanner.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/StoreScanner.html
index 581b976..498fef5 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/StoreScanner.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/StoreScanner.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":9,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":9,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -128,7 +128,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class StoreScanner
+public class StoreScanner
 extends NonReversedNonLazyKeyValueScanner
 implements KeyValueScanner, 
InternalScanner, 
ChangedReadersObserver
 Scanner scans both the memstore and the Store. Coalesce 
KeyValue stream
@@ -528,53 +528,54 @@ implements 
-protected ScanQueryMatcher.MatchCode
-optimize(ScanQueryMatcher.MatchCodeqcode,
-Cellcell)
-See if we should actually SEEK or rather just SKIP to the 
next Cell (see HBASE-13109).
-
-
-
 private void
 parallelSeek(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends KeyValueScannerscanners,
 Cellkv)
 Seek storefiles in parallel to optimize IO latency as much 
as possible
 
 
-
+
 Cell
 peek()
 Look at the next Cell in this scanner, but do not iterate 
scanner.
 
 
-
+
 boolean
 reseek(Cellkv)
 Reseek the scanner at or after the specified KeyValue.
 
 
-
+
 protected void
 resetKVHeap(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends KeyValueScannerscanners,
CellComparatorcomparator)
 
-
+
 protected void
 resetScannerStack(CelllastTopKey)
 
-
+
 boolean
 seek(Cellkey)
 Seek the scanner at or after the specified KeyValue.
 
 
-
+
 protected boolean
 seekAsDirection(Cellkv)
 Do a reseek in a normal StoreScanner(scan forward)
 
 
+
+private void
+seekOrSkipToNextColumn(Cellcell)
+
 
+private void
+seekOrSkipToNextRow(Cellcell)
+
+
 protected void
 seekScanners(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends KeyValueScannerscanners,
 CellseekKey,
@@ -583,24 +584,36 @@ implements Seek the specified scanners with the given key
 
 
-
+
 protected boolean
 seekToNextRow(Cellc)
 
-
+
 protected http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListKeyValueScanner
 selectScannersFrom(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">List? extends KeyValueScannerallScanners)
 Filters the given list of scanners using Bloom filter, time 
range, and
  TTL.
 
 
-
+
 void
 shipped()
 Called after a batch of rows scanned and set to be returned 
to client.
 
 
-
+
+protected boolean
+trySkipToNextColumn(Cellcell)
+See trySkipToNextRow(Cell)
+
+
+
+protected boolean
+trySkipToNextRow(Cellcell)
+See if we should actually SEEK or rather just SKIP to the 
next Cell (see HBASE-13109).
+
+
+
 void
 updateReaders(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListStoreFilesfs)
 Notify observers.
@@ -655,7 +668,7 @@ implements 
 
 LOG
-private static finalorg.apache.commons.logging.Log LOG
+private static finalorg.apache.commons.logging.Log LOG
 
 
 
@@ -664,7 +677,7 @@ implements 
 
 store
-protected finalStore store
+protected finalStore store
 
 
 
@@ -673,7 +686,7 @@ implements 
 
 matcher
-protectedScanQueryMatcher matcher
+protectedScanQueryMatcher matcher
 
 
 
@@ -682,7 +695,7 @@ implements 
 
 heap
-protectedKeyValueHeap heap
+protectedKeyValueHeap heap
 
 
 
@@ -691,7 +704,7 @@ implements 
 
 cacheBlocks
-protectedboolean cacheBlocks
+protectedboolean cacheBlocks
 
 
 
@@ -700,7 +713,7 @@ implements 
 
 countPerRow
-protectedlong countPerRow
+protectedlong countPerRow
 
 
 
@@ -709,7 +722,7 @@ implements 
 
 storeLimit
-protectedint storeLimit
+protectedint storeLimit
 
 
 
@@ -718,7 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-04-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6f2e75f2/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.Converter.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.Converter.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.Converter.html
index 6c52543..f3f7a46 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.Converter.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.Converter.html
@@ -31,1797 +31,2040 @@
 023import java.util.ArrayList;
 024import java.util.Arrays;
 025import java.util.Collection;
-026import java.util.HashMap;
-027import java.util.LinkedList;
-028import java.util.List;
-029import java.util.Map;
-030import java.util.Optional;
-031import 
java.util.concurrent.CompletableFuture;
-032import java.util.concurrent.TimeUnit;
-033import 
java.util.concurrent.atomic.AtomicReference;
-034import java.util.function.BiConsumer;
-035import java.util.regex.Pattern;
-036import java.util.stream.Collectors;
-037
-038import 
com.google.common.annotations.VisibleForTesting;
-039
-040import io.netty.util.Timeout;
-041import io.netty.util.TimerTask;
-042import org.apache.commons.logging.Log;
-043import 
org.apache.commons.logging.LogFactory;
-044import 
org.apache.hadoop.hbase.HColumnDescriptor;
-045import 
org.apache.hadoop.hbase.HRegionInfo;
-046import 
org.apache.hadoop.hbase.HRegionLocation;
-047import 
org.apache.hadoop.hbase.MetaTableAccessor;
-048import 
org.apache.hadoop.hbase.MetaTableAccessor.QueryType;
-049import 
org.apache.hadoop.hbase.NotServingRegionException;
-050import 
org.apache.hadoop.hbase.RegionLocations;
-051import 
org.apache.hadoop.hbase.ServerName;
-052import 
org.apache.hadoop.hbase.NamespaceDescriptor;
-053import 
org.apache.hadoop.hbase.HConstants;
-054import 
org.apache.hadoop.hbase.TableExistsException;
-055import 
org.apache.hadoop.hbase.TableName;
-056import 
org.apache.hadoop.hbase.AsyncMetaTableAccessor;
-057import 
org.apache.hadoop.hbase.TableNotFoundException;
-058import 
org.apache.hadoop.hbase.UnknownRegionException;
-059import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-060import 
org.apache.hadoop.hbase.classification.InterfaceStability;
-061import 
org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder;
-062import 
org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder;
-063import 
org.apache.hadoop.hbase.client.Scan.ReadType;
-064import 
org.apache.hadoop.hbase.client.replication.ReplicationSerDeHelper;
-065import 
org.apache.hadoop.hbase.client.replication.TableCFs;
-066import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-067import 
org.apache.hadoop.hbase.ipc.HBaseRpcController;
-068import 
org.apache.hadoop.hbase.quotas.QuotaFilter;
-069import 
org.apache.hadoop.hbase.quotas.QuotaSettings;
-070import 
org.apache.hadoop.hbase.quotas.QuotaTableUtil;
-071import 
org.apache.hadoop.hbase.replication.ReplicationException;
-072import 
org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
-073import 
org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
-074import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback;
-075import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-076import 
org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
-077import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
-078import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest;
-079import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse;
-080import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionRequest;
-081import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionResponse;
-082import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
-083import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema;
-084import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest;
-085import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse;
-086import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest;
-087import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse;
-088import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest;
-089import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse;
-090import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest;
-091import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse;
-092import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest;
-093import 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-04-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/efd0601e/checkstyle.rss
--
diff --git a/checkstyle.rss b/checkstyle.rss
index 3491926..1d3659b 100644
--- a/checkstyle.rss
+++ b/checkstyle.rss
@@ -25,8 +25,8 @@ under the License.
 en-us
 2007 - 2017 The Apache Software Foundation
 
-  File: 2146,
- Errors: 14236,
+  File: 2155,
+ Errors: 14362,
  Warnings: 0,
  Infos: 0
   
@@ -858,7 +858,7 @@ under the License.
   
   
 
-  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.replication.ReplicationLoadSource.java;>org/apache/hadoop/hbase/replication/ReplicationLoadSource.java
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.metrics.MetricRegistriesLoader.java;>org/apache/hadoop/hbase/metrics/MetricRegistriesLoader.java
 
 
   0
@@ -867,12 +867,12 @@ under the License.
   0
 
 
-  0
+  2
 
   
   
 
-  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.metrics.MetricRegistriesLoader.java;>org/apache/hadoop/hbase/metrics/MetricRegistriesLoader.java
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.replication.ReplicationLoadSource.java;>org/apache/hadoop/hbase/replication/ReplicationLoadSource.java
 
 
   0
@@ -881,7 +881,7 @@ under the License.
   0
 
 
-  2
+  0
 
   
   
@@ -1180,6 +1180,20 @@ under the License.
   
   
 
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.zookeeper.MetricsZooKeeper.java;>org/apache/hadoop/hbase/zookeeper/MetricsZooKeeper.java
+
+
+  0
+
+
+  0
+
+
+  3
+
+  
+  
+
   http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.util.Triple.java;>org/apache/hadoop/hbase/util/Triple.java
 
 
@@ -1390,7 +1404,7 @@ under the License.
   
   
 
-  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.ExportSnapshotException.java;>org/apache/hadoop/hbase/snapshot/ExportSnapshotException.java
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.replication.regionserver.MetricsReplicationSourceImpl.java;>org/apache/hadoop/hbase/replication/regionserver/MetricsReplicationSourceImpl.java
 
 
   0
@@ -1404,7 +1418,7 @@ under the License.
   
   
 
-  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.replication.regionserver.MetricsReplicationSourceImpl.java;>org/apache/hadoop/hbase/replication/regionserver/MetricsReplicationSourceImpl.java
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.ExportSnapshotException.java;>org/apache/hadoop/hbase/snapshot/ExportSnapshotException.java
 
 
   0
@@ -2071,7 +2085,7 @@ under the License.
   0
 
 
-  1
+  2
 
   
   
@@ -3420,7 +3434,7 @@ under the License.
   
   
 
-  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.client.AsyncConnectionConfiguration.java;>org/apache/hadoop/hbase/client/AsyncConnectionConfiguration.java
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.wal.MetricsEditsReplaySource.java;>org/apache/hadoop/hbase/regionserver/wal/MetricsEditsReplaySource.java
 
 
   0
@@ -3429,12 +3443,12 @@ under the License.
   0
 
 
-  1
+  0
 
   
   
 
-  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.wal.MetricsEditsReplaySource.java;>org/apache/hadoop/hbase/regionserver/wal/MetricsEditsReplaySource.java
+  

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-04-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/10601a30/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.AdminRpcCall.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.AdminRpcCall.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.AdminRpcCall.html
index be839b7..72853dd 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.AdminRpcCall.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.AdminRpcCall.html
@@ -45,1639 +45,1784 @@
 037
 038import 
com.google.common.annotations.VisibleForTesting;
 039
-040import org.apache.commons.logging.Log;
-041import 
org.apache.commons.logging.LogFactory;
-042import 
org.apache.hadoop.hbase.HColumnDescriptor;
-043import 
org.apache.hadoop.hbase.HRegionInfo;
-044import 
org.apache.hadoop.hbase.HRegionLocation;
-045import 
org.apache.hadoop.hbase.HTableDescriptor;
-046import 
org.apache.hadoop.hbase.MetaTableAccessor;
-047import 
org.apache.hadoop.hbase.MetaTableAccessor.QueryType;
-048import 
org.apache.hadoop.hbase.NotServingRegionException;
-049import 
org.apache.hadoop.hbase.RegionLocations;
-050import 
org.apache.hadoop.hbase.ServerName;
-051import 
org.apache.hadoop.hbase.NamespaceDescriptor;
-052import 
org.apache.hadoop.hbase.HConstants;
-053import 
org.apache.hadoop.hbase.TableName;
-054import 
org.apache.hadoop.hbase.AsyncMetaTableAccessor;
-055import 
org.apache.hadoop.hbase.TableNotFoundException;
-056import 
org.apache.hadoop.hbase.UnknownRegionException;
-057import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-058import 
org.apache.hadoop.hbase.classification.InterfaceStability;
-059import 
org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder;
-060import 
org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder;
-061import 
org.apache.hadoop.hbase.client.Scan.ReadType;
-062import 
org.apache.hadoop.hbase.client.replication.ReplicationSerDeHelper;
-063import 
org.apache.hadoop.hbase.client.replication.TableCFs;
-064import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-065import 
org.apache.hadoop.hbase.ipc.HBaseRpcController;
-066import 
org.apache.hadoop.hbase.quotas.QuotaFilter;
-067import 
org.apache.hadoop.hbase.quotas.QuotaSettings;
-068import 
org.apache.hadoop.hbase.quotas.QuotaTableUtil;
-069import 
org.apache.hadoop.hbase.replication.ReplicationException;
-070import 
org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
-071import 
org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
-072import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback;
-073import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-074import 
org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
-075import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
-076import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest;
-077import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse;
-078import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionRequest;
-079import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionResponse;
-080import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema;
-081import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest;
-082import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse;
-083import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest;
-084import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse;
-085import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest;
-086import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse;
-087import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest;
-088import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse;
-089import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest;
-090import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse;
-091import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest;
-092import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse;
-093import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest;
-094import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse;
-095import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest;
-096import 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-04-20 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/662ea7dc/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.CreateTableProcedureBiConsumer.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.CreateTableProcedureBiConsumer.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.CreateTableProcedureBiConsumer.html
index ac4a9b3..be839b7 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.CreateTableProcedureBiConsumer.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.CreateTableProcedureBiConsumer.html
@@ -30,212 +30,212 @@
 022import java.io.IOException;
 023import java.util.ArrayList;
 024import java.util.Arrays;
-025import java.util.LinkedList;
-026import java.util.List;
-027import java.util.Optional;
-028import 
java.util.concurrent.CompletableFuture;
-029import java.util.concurrent.TimeUnit;
-030import 
java.util.concurrent.atomic.AtomicReference;
-031import java.util.function.BiConsumer;
-032import java.util.regex.Pattern;
-033
-034import 
com.google.common.annotations.VisibleForTesting;
-035import org.apache.commons.logging.Log;
-036import 
org.apache.commons.logging.LogFactory;
-037import 
org.apache.hadoop.hbase.HColumnDescriptor;
-038import 
org.apache.hadoop.hbase.HRegionInfo;
-039import 
org.apache.hadoop.hbase.HRegionLocation;
-040import 
org.apache.hadoop.hbase.HTableDescriptor;
-041import 
org.apache.hadoop.hbase.MetaTableAccessor;
-042import 
org.apache.hadoop.hbase.MetaTableAccessor.QueryType;
-043import 
org.apache.hadoop.hbase.NotServingRegionException;
-044import 
org.apache.hadoop.hbase.RegionLocations;
-045import 
org.apache.hadoop.hbase.ServerName;
-046import 
org.apache.hadoop.hbase.NamespaceDescriptor;
-047import 
org.apache.hadoop.hbase.HConstants;
-048import 
org.apache.hadoop.hbase.TableName;
-049import 
org.apache.hadoop.hbase.AsyncMetaTableAccessor;
-050import 
org.apache.hadoop.hbase.TableNotFoundException;
-051import 
org.apache.hadoop.hbase.UnknownRegionException;
-052import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-053import 
org.apache.hadoop.hbase.classification.InterfaceStability;
-054import 
org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder;
-055import 
org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder;
-056import 
org.apache.hadoop.hbase.client.Scan.ReadType;
-057import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-058import 
org.apache.hadoop.hbase.ipc.HBaseRpcController;
-059import 
org.apache.hadoop.hbase.quotas.QuotaFilter;
-060import 
org.apache.hadoop.hbase.quotas.QuotaSettings;
-061import 
org.apache.hadoop.hbase.quotas.QuotaTableUtil;
-062import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback;
-063import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-064import 
org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
-065import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
-066import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest;
-067import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse;
-068import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionRequest;
-069import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionResponse;
-070import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema;
-071import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest;
-072import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse;
-073import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest;
-074import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse;
-075import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest;
-076import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse;
-077import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest;
-078import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse;
-079import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest;
-080import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse;
-081import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest;
-082import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse;
-083import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest;
-084import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse;
-085import 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-04-19 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6b4bae59/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html 
b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
index b9941b0..cea7361 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
@@ -1308,7 +1308,7 @@ implements HRegionServer
-abort,
 addToMovedRegions,
 addToOnlineRegions,
 checkFileSystem,
 cleanMovedRegions,
 closeAllRegions,
 closeAndOfflineRegion
 ForSplitOrMerge, closeRegion,
 constructRegionServer,
 convertThrowableToIOE,
 createClusterConnection,
 createRegionLoad,
 createRegionServerS
 tatusStub, createRegionServerStatusStub,
 execRegionServerService,
 getCacheConfig,
 getChoreService,
 getClusterConnection,
 getClusterId,
 getCompactionPressure, getCompactionRequester,
 getCompactSplitThread,
 getConfiguration,
 getConfigurationManager,
 getConnection,
 getCoordinatedStateManager,
 getExecutorService,
 getFavoredNodesForRegion,
 getFileSystem,
 getFlushPressure,
 getFlushRequester,
 getFlushThroughputController,
 getFromOnlineRegions,
 getHeapMemoryManager,
 getInfoServer, getLastSequenceId,
 getLeases,
 getMasterAddressTracker,
 getMetaTableLocator,
 getMetrics,
 getMostLoadedRegions,
 getNonceManager,
 getNumberOfOnlineRegions, getOnlineRegion,
 getOnlineRegions,
 getOnlineRegions,
 getOnlineRegionsLocalContext,
 getOnlineTables,
 getRecoveringRegions,
 getRegion,
 getRegionBlockLocations,
 getRegionByEncodedName,
 getRegionByEncodedName,
 getRegionServerAccounting,
 getRegionServerCoprocessorHost,
 getRegionServerCoprocessors,
 getRegionServerMetrics,
 getRegionServerQuotaManager,
 getRegionsInTransitionInRS,
 getReplicationSourceService,
 getRootDir,
 getRpcServer,
 getRSRpcServices,
 getSecureBulkLoadManager,
 getStartcode, getThreadWakeFrequency,
 getWAL,
 getWALFileSystem,
 getWalRoller,
 getWALRootDir,
 getWALs,
 handleReportForDutyResponse,
 isAborted,
 isOnline,
 isProcedureFinished,
 isStopped,
 isStopping,
 kill,
 movedRegionCleanerPeriod,
 onConfigurationChange,
 postOpenDeployTasks,
 postOpenDeployTasks,
 regionLock,
 removeFromOnlineRegions,
 reportRegionStateTransition,
 reportRegionStateTransition,
 reportRegionStateTransition,
 requestRegionSplit,
 run,
 setInitLatch,
 setupClusterConnection,
 shouldUseThisHostnameInstead,
 stop,
 toString,
 tryRegionServerReport,
 unassign,
 updateConfiguration,
 updateRegionFavoredNodesMapping, waitForServerOnline,
 walRollRequestFinished
+abort,
 addToMovedRegions,
 addToOnlineRegions,
 checkFileSystem,
 cleanMovedRegions,
 closeAllRegions,
 closeAndOfflineRegion
 ForSplitOrMerge, closeRegion,
 constructRegionServer,
 convertThrowableToIOE,
 createClusterConnection,
 createRegionLoad,
 createRegionServerS
 tatusStub, createRegionServerStatusStub,
 execRegionServerService,
 getCacheConfig,
 getChoreService,
 getClusterConnection,
 getClusterId,
 getCompactionPressure, getCompactionRequester,
 getCompactSplitThread,
 getConfiguration,
 getConfigurationManager,
 getConnection,
 getCoordinatedStateManager,
 getExecutorService,
 getFavoredNodesForRegion,
 getFileSystem,
 getFlushPressure,
 getFlushRequester,
 getFlushThroughputController,
 getFromOnlineRegions,
 getHeapMemoryManager,
 getInfoServer, getLastSequenceId,
 getLeases,
 getMasterAddressTracker,
 getMetaTableLocator,
 getMetrics,
 getMostLoadedRegions,
 getNonceManager,
 getNumberOfOnlineRegions, getOnlineRegion,
 getOnlineRegions,
 getOnlineRegions,
 getOnlineRegionsLocalContext,
 getOnlineTables,
 getRecoveringRegions,
 getRegion,
 getRegionBlockLocations,
 getRegionByEncodedName,
 getRegionByEncodedName,
 getRegionServerAccounting,
 getRegionServerCoprocessorHost,
 getRegionServerCoprocessors,
 getRegionServerMetrics,
 getRegionServerQuotaManager,
 getRegionsInTransitionInRS,
 getReplicationSourceService,
 getRootDir,
 getRpcServer,
 getRSRpcServices,
 getSecureBulkLoadManager,
 getStartcode, getThreadWakeFrequency,
 getWAL,
 getWALFileSystem,
 getWalRoller,
 getWALRootDir,
 getWALs,
 handleReportForDutyResponse,
 initializeMemStoreChunkCreator,
 isAborted,
 isOnline,
 isProcedureFinished,
 isStopped,
 isStopping,
 kill,
 movedRegionCleanerPeriod,
 onConfigurationChange,
 postOpenDeployTasks,
 postOpenDeployTasks,
 regionLock,
 removeFromOnlineRegions,
 reportRegionStateTransition,
 reportRegionStateTransition,
 reportRegionStateTransition,
 requestRegionSplit,
 run,
 setInitLatch,
 setupClusterConnection,
 shouldUseThisHostnameInstead,
 stop,
 toString,
 tryRegionServerReport,
 unassign,
 updateConfigu
 ration, updateRegionFavoredNodesMapping,
 waitForServerOnline,
 walRollRequestFinished
 
 
 
@@ -2185,7 +2185,7 @@ implements 
 
 initMobCleaner
-privatevoidinitMobCleaner()
+privatevoidinitMobCleaner()
 
 
 
@@ -2194,7 +2194,7 @@ implements 
 
 createMetaBootstrap

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-04-18 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2fcc2ae0/devapidocs/org/apache/hadoop/hbase/tool/Canary.RegionServerMonitor.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/tool/Canary.RegionServerMonitor.html 
b/devapidocs/org/apache/hadoop/hbase/tool/Canary.RegionServerMonitor.html
index 4b815e3..1a782ff 100644
--- a/devapidocs/org/apache/hadoop/hbase/tool/Canary.RegionServerMonitor.html
+++ b/devapidocs/org/apache/hadoop/hbase/tool/Canary.RegionServerMonitor.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static class Canary.RegionServerMonitor
+private static class Canary.RegionServerMonitor
 extends Canary.Monitor
 
 
@@ -254,7 +254,7 @@ extends 
 
 allRegions
-privateboolean allRegions
+privateboolean allRegions
 
 
 
@@ -271,7 +271,7 @@ extends 
 
 RegionServerMonitor
-publicRegionServerMonitor(Connectionconnection,
+publicRegionServerMonitor(Connectionconnection,
http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String[]monitorTargets,
booleanuseRegExp,
Canary.ExtendedSinksink,
@@ -294,7 +294,7 @@ extends 
 
 getSink
-privateCanary.ExtendedSinkgetSink()
+privateCanary.ExtendedSinkgetSink()
 
 
 
@@ -303,7 +303,7 @@ extends 
 
 run
-publicvoidrun()
+publicvoidrun()
 
 Specified by:
 http://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true#run--;
 title="class or interface in java.lang">runin 
interfacehttp://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true;
 title="class or interface in java.lang">Runnable
@@ -318,7 +318,7 @@ extends 
 
 checkNoTableNames
-privatebooleancheckNoTableNames()
+privatebooleancheckNoTableNames()
 
 
 
@@ -327,7 +327,7 @@ extends 
 
 monitorRegionServers
-privatevoidmonitorRegionServers(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListHRegionInforsAndRMap)
+privatevoidmonitorRegionServers(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListHRegionInforsAndRMap)
 
 
 
@@ -336,7 +336,7 @@ extends 
 
 filterRegionServerByName
-privatehttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListHRegionInfofilterRegionServerByName()
+privatehttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListHRegionInfofilterRegionServerByName()
 
 
 
@@ -345,7 +345,7 @@ extends 
 
 getAllRegionServerByName
-privatehttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListHRegionInfogetAllRegionServerByName()
+privatehttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListHRegionInfogetAllRegionServerByName()
 
 
 
@@ -354,7 +354,7 @@ extends 
 
 doFilterRegionServerByName
-privatehttp://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-04-17 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e4348f53/devapidocs/org/apache/hadoop/hbase/regionserver/BusyRegionSplitPolicy.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/BusyRegionSplitPolicy.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/BusyRegionSplitPolicy.html
index 9fa46dc..5bc3c7f 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/BusyRegionSplitPolicy.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/BusyRegionSplitPolicy.html
@@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab";
 
 
 PrevClass
-NextClass
+NextClass
 
 
 Frames
@@ -563,7 +563,7 @@ extends 
 
 PrevClass
-NextClass
+NextClass
 
 
 Frames

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e4348f53/devapidocs/org/apache/hadoop/hbase/regionserver/ByteBufferChunkCell.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/ByteBufferChunkCell.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/ByteBufferChunkCell.html
new file mode 100644
index 000..f08b4b2
--- /dev/null
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/ByteBufferChunkCell.html
@@ -0,0 +1,353 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+
+ByteBufferChunkCell (Apache HBase 2.0.0-SNAPSHOT API)
+
+
+
+
+
+var methods = {"i0":10};
+var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
+var altColor = "altColor";
+var rowColor = "rowColor";
+var tableTab = "tableTab";
+var activeTableTab = "activeTableTab";
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+PrevClass
+NextClass
+
+
+Frames
+NoFrames
+
+
+AllClasses
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.hadoop.hbase.regionserver
+Class 
ByteBufferChunkCell
+
+
+
+http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">java.lang.Object
+
+
+org.apache.hadoop.hbase.ByteBufferCell
+
+
+org.apache.hadoop.hbase.ByteBufferKeyValue
+
+
+org.apache.hadoop.hbase.regionserver.ByteBufferChunkCell
+
+
+
+
+
+
+
+
+
+
+
+All Implemented Interfaces:
+http://docs.oracle.com/javase/8/docs/api/java/lang/Cloneable.html?is-external=true;
 title="class or interface in java.lang">Cloneable, Cell, ExtendedCell, HeapSize, SettableSequenceId, SettableTimestamp
+
+
+
+@InterfaceAudience.Private
+public class ByteBufferChunkCell
+extends ByteBufferKeyValue
+ByteBuffer based cell which has the chunkid at the 0th 
offset
+
+See Also:
+MemStoreLAB
+
+
+
+
+
+
+
+
+
+
+
+
+Field Summary
+
+
+
+
+Fields inherited from classorg.apache.hadoop.hbase.ByteBufferKeyValue
+buf,
 FIXED_OVERHEAD,
 length,
 offset
+
+
+
+
+
+Fields inherited from interfaceorg.apache.hadoop.hbase.ExtendedCell
+CELL_NOT_BASED_ON_CHUNK
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+ByteBufferChunkCell(http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
 title="class or interface in java.nio">ByteBufferbuf,
+   intoffset,
+   intlength)
+
+
+ByteBufferChunkCell(http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
 title="class or interface in java.nio">ByteBufferbuf,
+   intoffset,
+   intlength,
+   longseqId)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+All MethodsInstance MethodsConcrete Methods
+
+Modifier and Type
+Method and Description
+
+
+int
+getChunkId()
+Extracts the id of the backing bytebuffer of this cell if 
it was obtained from fixed sized
+ chunks as in case of MemstoreLAB
+
+
+
+
+
+
+
+Methods inherited from classorg.apache.hadoop.hbase.ByteBufferKeyValue
+deepClone,
 equals,
 getBuffer,
 getFamilyArray,
 getFamilyByteBuffer,
 getFamilyLength,
 getFamilyOffset,
 getFamilyPosition,
 getOffset,
 getQualifierArray,
 getQualifierByteBuffer,
 getQualifierLength,
 getQualifierOffset,
 getQualifierPosition,
 getRowArray,
 getRowByteBuffer,
 getRowLength,
 getRowOffset, getRowPosition,
 getSequenceId,
 getSerializedSize,
 getTagsArray,
 getTagsByteBuffer,
 getTagsLength,
 getTagsOffset,
 getTagsPosition,
 getTimestamp, getTypeByte,
 getValueArray,
 getValueByteBuffer,
 getValueLength,
 getValueOffset,
 getValuePosition,
 hashCode,
 heapSize,
 

[28/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-04-11 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e57d1b63/apidocs/org/apache/hadoop/hbase/TableInfoMissingException.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/TableInfoMissingException.html 
b/apidocs/org/apache/hadoop/hbase/TableInfoMissingException.html
deleted file mode 100644
index f13e6ec..000
--- a/apidocs/org/apache/hadoop/hbase/TableInfoMissingException.html
+++ /dev/null
@@ -1,313 +0,0 @@
-http://www.w3.org/TR/html4/loose.dtd;>
-
-
-
-
-
-TableInfoMissingException (Apache HBase 2.0.0-SNAPSHOT API)
-
-
-
-
-
-
-
-JavaScript is disabled on your browser.
-
-
-
-
-
-Skip navigation links
-
-
-
-
-Overview
-Package
-Class
-Use
-Tree
-Deprecated
-Index
-Help
-
-
-
-
-PrevClass
-NextClass
-
-
-Frames
-NoFrames
-
-
-AllClasses
-
-
-
-
-
-
-
-Summary:
-Nested|
-Field|
-Constr|
-Method
-
-
-Detail:
-Field|
-Constr|
-Method
-
-
-
-
-
-
-
-
-org.apache.hadoop.hbase
-Class 
TableInfoMissingException
-
-
-
-http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">java.lang.Object
-
-
-http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
 title="class or interface in java.lang">java.lang.Throwable
-
-
-http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
 title="class or interface in java.lang">java.lang.Exception
-
-
-http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">java.io.IOException
-
-
-org.apache.hadoop.hbase.HBaseIOException
-
-
-org.apache.hadoop.hbase.TableInfoMissingException
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-All Implemented Interfaces:
-http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable
-
-
-
-@InterfaceAudience.Public
-public class TableInfoMissingException
-extends HBaseIOException
-Failed to find .tableinfo file under table dir
-
-See Also:
-Serialized
 Form
-
-
-
-
-
-
-
-
-
-
-
-
-Constructor Summary
-
-Constructors
-
-Constructor and Description
-
-
-TableInfoMissingException()
-
-
-TableInfoMissingException(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">Stringmessage)
-
-
-TableInfoMissingException(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringmessage,
- http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
 title="class or interface in java.lang">Throwablet)
-
-
-TableInfoMissingException(http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
 title="class or interface in java.lang">Throwablet)
-
-
-
-
-
-
-
-
-
-Method Summary
-
-
-
-
-Methods inherited from classjava.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
 title="class or interface in java.lang">Throwable
-http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#addSuppressed-java.lang.Throwable-;
 title="class or interface in java.lang">addSuppressed, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#fillInStackTrace--;
 title="class or interface in java.lang">fillInStackTrace, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getCause--;
 title="class or interface in java.lang">getCause, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getLocalizedMessage--;
 title="class or interface in java.lang">getLocalizedMessage, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getMessage--;
 title="class or interface in java.lang">getMessage, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getStackTrace--;
 title="class or inter
 face in java.lang">getStackTrace, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getSuppressed--;
 title="class or interface in java.lang">getSuppressed, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#initCause-java.lang.Throwable-;
 title="class or interface in java.lang">initCause, http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#printStackTrace--;
 title="class or interface in java.lang">printStackTrace,