[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-08-11 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/346adc37/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
index cbc2ae9..9f4d04d 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html
@@ -692,20 +692,20 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
 title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage
+org.apache.hadoop.hbase.regionserver.MemStoreCompactor.Action
+org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection
 org.apache.hadoop.hbase.regionserver.FlushType
-org.apache.hadoop.hbase.regionserver.ScannerContext.NextState
-org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status
-org.apache.hadoop.hbase.regionserver.StoreScanner.StoreScannerCompactionRace
-org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType
-org.apache.hadoop.hbase.regionserver.BloomType
 org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope
 org.apache.hadoop.hbase.regionserver.Region.Operation
+org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType
 org.apache.hadoop.hbase.regionserver.ScanType
-org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection
-org.apache.hadoop.hbase.regionserver.Region.FlushResult.Result
-org.apache.hadoop.hbase.regionserver.MemStoreCompactor.Action
 org.apache.hadoop.hbase.regionserver.RegionOpeningState
+org.apache.hadoop.hbase.regionserver.BloomType
+org.apache.hadoop.hbase.regionserver.ScannerContext.NextState
+org.apache.hadoop.hbase.regionserver.StoreScanner.StoreScannerCompactionRace
+org.apache.hadoop.hbase.regionserver.Region.FlushResult.Result
+org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status
+org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/346adc37/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/NormalUserScanQueryMatcher.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/NormalUserScanQueryMatcher.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/NormalUserScanQueryMatcher.html
index e8361b7..f1461a6 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/NormalUserScanQueryMatcher.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/NormalUserScanQueryMatcher.html
@@ -380,7 +380,7 @@ extends Specified by:
 beforeShippedin
 interfaceShipperListener
 Overrides:
-beforeShippedin
 classScanQueryMatcher
+beforeShippedin
 classUserScanQueryMatcher
 Throws:
 http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
 title="class or interface in java.io">IOException
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/346adc37/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/RawScanQueryMatcher.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/RawScanQueryMatcher.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/RawScanQueryMatcher.html
index cb4af05..307e792 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/RawScanQueryMatcher.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/RawScanQueryMatcher.html
@@ -242,14 +242,14 @@ extends UserScanQueryMatcher
-create,
 getFilter,
 getNextKeyHint,
 hasNullColumnInQuery,
 isUserScan,
 matchColumn, moreRowsMayExistAfter,
 moreRowsMayExistsAfter
+beforeShipped,
 create,
 getFilter,
 getNextKeyHint,
 hasNullColumnInQuery,
 isUserScan, matchColumn,
 moreRowsMayExistAfter,
 moreRowsMayExistsAfter
 
 
 
 
 
 Methods inherited from 
classorg.apache.hadoop.hbase.regionserver.querymatcher.ScanQueryMatcher
-beforeShipped,
 checkColumn,
 checkDeleted,
 clearCurrentRow,
 compareKeyForNextColumn,
 compareKeyForNextRow,
 createStartKeyFromRow,
 currentRow,
 getKeyForNextColumn,
 getStartKey,
 getTrackers,
 preCheck,
 setToNewRow
+checkColumn,
 checkDeleted,
 clearCurrentRow,
 compareKeyForNextColumn,
 compareK
 eyForNextRow, createStartKeyFromRow,
 currentRow,
 getKeyForNextColumn,
 getStartKey,
 getTrackers,
 preCheck,
 

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-08-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2aec596e/devapidocs/org/apache/hadoop/hbase/client/Admin.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/Admin.html 
b/devapidocs/org/apache/hadoop/hbase/client/Admin.html
index 281158f..acc88a2 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/Admin.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/Admin.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":6,"i1":6,"i2":6,"i3":38,"i4":6,"i5":6,"i6":18,"i7":18,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6,"i20":6,"i21":6,"i22":6,"i23":6,"i24":6,"i25":6,"i26":6,"i27":6,"i28":6,"i29":6,"i30":6,"i31":6,"i32":50,"i33":50,"i34":50,"i35":6,"i36":6,"i37":6,"i38":50,"i39":6,"i40":38,"i41":6,"i42":6,"i43":6,"i44":6,"i45":6,"i46":6,"i47":6,"i48":6,"i49":6,"i50":6,"i51":38,"i52":38,"i53":6,"i54":6,"i55":18,"i56":6,"i57":6,"i58":6,"i59":38,"i60":38,"i61":6,"i62":6,"i63":18,"i64":6,"i65":6,"i66":6,"i67":38,"i68":38,"i69":6,"i70":6,"i71":6,"i72":6,"i73":38,"i74":6,"i75":6,"i76":6,"i77":6,"i78":6,"i79":6,"i80":6,"i81":6,"i82":6,"i83":6,"i84":6,"i85":6,"i86":6,"i87":6,"i88":6,"i89":6,"i90":6,"i91":18,"i92":6,"i93":38,"i94":38,"i95":38,"i96":6,"i97":6,"i98":6,"i99":6,"i100":6,"i101":6,"i102":6,"i103":6,"i104":6,"i105":6,"i106":6,"i107":6,"i108":6,"i109":6,"i110":6,"i111":6,"i112":6,"i113":6,"i114":6,"i115":18,"i116":18,"i117":18,"i118"
 
:6,"i119":6,"i120":6,"i121":6,"i122":6,"i123":6,"i124":6,"i125":6,"i126":6,"i127":6,"i128":6,"i129":38,"i130":6,"i131":6,"i132":6,"i133":6,"i134":6,"i135":6,"i136":38,"i137":38,"i138":38,"i139":38,"i140":38,"i141":6,"i142":6,"i143":6,"i144":6,"i145":6,"i146":6,"i147":6,"i148":6,"i149":38,"i150":6,"i151":6,"i152":38,"i153":6,"i154":6,"i155":6,"i156":6,"i157":6,"i158":38,"i159":6,"i160":38,"i161":6,"i162":6,"i163":6,"i164":6,"i165":18,"i166":18,"i167":6,"i168":6,"i169":6,"i170":6,"i171":6,"i172":6,"i173":6,"i174":6,"i175":6,"i176":6,"i177":6,"i178":6,"i179":6,"i180":6,"i181":6,"i182":6,"i183":6,"i184":6,"i185":6,"i186":6,"i187":6,"i188":38,"i189":38,"i190":6,"i191":6,"i192":6,"i193":6,"i194":6,"i195":6,"i196":6,"i197":6,"i198":6,"i199":6,"i200":18};
+var methods = 
{"i0":6,"i1":6,"i2":6,"i3":38,"i4":6,"i5":6,"i6":18,"i7":18,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":38,"i18":38,"i19":38,"i20":38,"i21":6,"i22":6,"i23":6,"i24":6,"i25":6,"i26":6,"i27":6,"i28":6,"i29":6,"i30":6,"i31":6,"i32":50,"i33":50,"i34":50,"i35":6,"i36":6,"i37":6,"i38":50,"i39":6,"i40":38,"i41":6,"i42":6,"i43":6,"i44":6,"i45":6,"i46":6,"i47":6,"i48":6,"i49":6,"i50":6,"i51":38,"i52":38,"i53":6,"i54":6,"i55":18,"i56":6,"i57":6,"i58":6,"i59":38,"i60":38,"i61":6,"i62":6,"i63":18,"i64":6,"i65":6,"i66":6,"i67":38,"i68":38,"i69":6,"i70":6,"i71":6,"i72":6,"i73":38,"i74":6,"i75":6,"i76":6,"i77":6,"i78":6,"i79":6,"i80":6,"i81":6,"i82":6,"i83":6,"i84":6,"i85":6,"i86":6,"i87":6,"i88":6,"i89":6,"i90":6,"i91":18,"i92":6,"i93":38,"i94":38,"i95":38,"i96":6,"i97":6,"i98":6,"i99":6,"i100":6,"i101":6,"i102":6,"i103":6,"i104":6,"i105":6,"i106":6,"i107":6,"i108":6,"i109":6,"i110":6,"i111":6,"i112":6,"i113":6,"i114":6,"i115":18,"i116":18,"i117":18,"i
 
118":6,"i119":6,"i120":6,"i121":6,"i122":6,"i123":6,"i124":6,"i125":6,"i126":6,"i127":6,"i128":6,"i129":38,"i130":6,"i131":6,"i132":6,"i133":6,"i134":6,"i135":6,"i136":38,"i137":38,"i138":38,"i139":38,"i140":38,"i141":6,"i142":6,"i143":6,"i144":6,"i145":6,"i146":6,"i147":6,"i148":6,"i149":38,"i150":6,"i151":6,"i152":38,"i153":6,"i154":6,"i155":6,"i156":6,"i157":6,"i158":38,"i159":6,"i160":38,"i161":6,"i162":6,"i163":6,"i164":6,"i165":18,"i166":18,"i167":6,"i168":6,"i169":6,"i170":6,"i171":6,"i172":6,"i173":6,"i174":6,"i175":6,"i176":6,"i177":6,"i178":6,"i179":6,"i180":6,"i181":6,"i182":6,"i183":6,"i184":6,"i185":6,"i186":6,"i187":6,"i188":38,"i189":38,"i190":6,"i191":6,"i192":6,"i193":6,"i194":6,"i195":6,"i196":6,"i197":6,"i198":6,"i199":6,"i200":18};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],4:["t3","Abstract Methods"],16:["t5","Default 
Methods"],32:["t6","Deprecated Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -264,28 +264,36 @@ extends void
 closeRegion(byte[]regionname,
http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringserverName)
-Close a region.
+Deprecated.
+Since 2.0. Will be removed 
in 3.0. Use unassign(byte[],
 boolean) instead.
+
 
 
 
 void
 closeRegion(ServerNamesn,
HRegionInfohri)
-Close a region.
+Deprecated.
+Since 2.0. Will be removed 
in 3.0. Use unassign(byte[],
 boolean) instead.
+
 
 
 
 void
 closeRegion(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringregionname,

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-31 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1837997e/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/AnnotationReadingPriorityFunction.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/AnnotationReadingPriorityFunction.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/AnnotationReadingPriorityFunction.html
index 7e411c6..39d8ea2 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/AnnotationReadingPriorityFunction.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/AnnotationReadingPriorityFunction.html
@@ -41,255 +41,253 @@
 033import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest;
 034import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest;
 035import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest;
-036import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionRequest;
-037import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.GetRequest;
-038import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.MutateRequest;
-039import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ScanRequest;
-040import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier;
-041import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
-042
-043import 
org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
-044import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
-045import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat;
-046import 
org.apache.hadoop.hbase.security.User;
-047
-048/**
-049 * Reads special method annotations and 
table names to figure a priority for use by QoS facility in
-050 * ipc; e.g: rpcs to hbase:meta get 
priority.
-051 */
-052// TODO: Remove.  This is doing way too 
much work just to figure a priority.  Do as Elliott
-053// suggests and just have the client 
specify a priority.
-054
-055//The logic for figuring out high 
priority RPCs is as follows:
-056//1. if the method is annotated with a 
QosPriority of QOS_HIGH,
-057//   that is honored
-058//2. parse out the protobuf message and 
see if the request is for meta
-059//   region, and if so, treat it as a 
high priority RPC
-060//Some optimizations for (2) are done 
here -
-061//Clients send the argument classname as 
part of making the RPC. The server
-062//decides whether to deserialize the 
proto argument message based on the
-063//pre-established set of argument classes 
(knownArgumentClasses below).
-064//This prevents the server from having to 
deserialize all proto argument
-065//messages prematurely.
-066//All the argument classes declare a 
'getRegion' method that returns a
-067//RegionSpecifier object. Methods can be 
invoked on the returned object
-068//to figure out whether it is a meta 
region or not.
-069@InterfaceAudience.Private
-070public class 
AnnotationReadingPriorityFunction implements PriorityFunction {
-071  private static final Log LOG =
-072
LogFactory.getLog(AnnotationReadingPriorityFunction.class.getName());
-073
-074  /** Used to control the scan delay, 
currently sqrt(numNextCall * weight) */
-075  public static final String 
SCAN_VTIME_WEIGHT_CONF_KEY = "hbase.ipc.server.scan.vtime.weight";
-076
-077  protected final MapString, 
Integer annotatedQos;
-078  //We need to mock the regionserver 
instance for some unit tests (set via
-079  //setRegionServer method.
-080  private RSRpcServices rpcServices;
-081  @SuppressWarnings("unchecked")
-082  private final Class? extends 
Message[] knownArgumentClasses = new Class[]{
-083  GetRegionInfoRequest.class,
-084  GetStoreFileRequest.class,
-085  CloseRegionRequest.class,
-086  FlushRegionRequest.class,
-087  SplitRegionRequest.class,
-088  CompactRegionRequest.class,
-089  GetRequest.class,
-090  MutateRequest.class,
-091  ScanRequest.class
-092  };
-093
-094  // Some caches for helping 
performance
-095  private final MapString, Class? 
extends Message argumentToClassMap = new HashMap();
-096  private final MapString, 
MapClass? extends Message, Method methodMap = new 
HashMap();
+036import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.GetRequest;
+037import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.MutateRequest;
+038import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ScanRequest;
+039import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier;
+040import 
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader;
+041
+042import 
org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
+043import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
+044import 

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-28 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a2b2dd19/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapperImpl.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapperImpl.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapperImpl.html
index 80168c0..0066ffd 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapperImpl.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapperImpl.html
@@ -60,205 +60,237 @@
 052  private long minStoreFileAge;
 053  private long avgStoreFileAge;
 054  private long numReferenceFiles;
-055
-056  private ScheduledFuture? 
regionMetricsUpdateTask;
+055  private long maxFlushQueueSize;
+056  private long maxCompactionQueueSize;
 057
-058  public MetricsRegionWrapperImpl(HRegion 
region) {
-059this.region = region;
-060this.executor = 
CompatibilitySingletonFactory.getInstance(MetricsExecutor.class).getExecutor();
-061this.runnable = new 
HRegionMetricsWrapperRunnable();
-062this.regionMetricsUpdateTask = 
this.executor.scheduleWithFixedDelay(this.runnable, PERIOD,
-063  PERIOD, TimeUnit.SECONDS);
-064  }
-065
-066  @Override
-067  public String getTableName() {
-068TableDescriptor tableDesc = 
this.region.getTableDescriptor();
-069if (tableDesc == null) {
-070  return UNKNOWN;
-071}
-072return 
tableDesc.getTableName().getQualifierAsString();
-073  }
-074
-075  @Override
-076  public String getNamespace() {
-077TableDescriptor tableDesc = 
this.region.getTableDescriptor();
-078if (tableDesc == null) {
-079  return UNKNOWN;
-080}
-081return 
tableDesc.getTableName().getNamespaceAsString();
-082  }
-083
-084
-085  @Override
-086  public String getRegionName() {
-087HRegionInfo regionInfo = 
this.region.getRegionInfo();
-088if (regionInfo == null) {
-089  return UNKNOWN;
-090}
-091return regionInfo.getEncodedName();
-092  }
-093
-094  @Override
-095  public long getNumStores() {
-096Mapbyte[],Store stores = 
this.region.stores;
-097if (stores == null) {
-098  return 0;
-099}
-100return stores.size();
-101  }
-102
-103  @Override
-104  public long getNumStoreFiles() {
-105return numStoreFiles;
-106  }
-107
-108  @Override
-109  public long getMemstoreSize() {
-110return memstoreSize;
-111  }
-112
-113  @Override
-114  public long getStoreFileSize() {
-115return storeFileSize;
-116  }
-117
-118  @Override
-119  public long getReadRequestCount() {
-120return 
this.region.getReadRequestsCount();
-121  }
-122
-123  @Override
-124  public long 
getFilteredReadRequestCount() {
-125return 
this.region.getFilteredReadRequestsCount();
-126  }
-127
-128  @Override
-129  public long getWriteRequestCount() {
-130return 
this.region.getWriteRequestsCount();
-131  }
-132
-133  @Override
-134  public long getNumFilesCompacted() {
-135return 
this.region.compactionNumFilesCompacted.get();
-136  }
-137
-138  @Override
-139  public long getNumBytesCompacted() {
-140return 
this.region.compactionNumBytesCompacted.get();
-141  }
-142
-143  @Override
-144  public long 
getNumCompactionsCompleted() {
-145return 
this.region.compactionsFinished.get();
-146  }
-147
-148  @Override
-149  public long getLastMajorCompactionAge() 
{
-150long lastMajorCompactionTs = 0L;
-151try {
-152  lastMajorCompactionTs = 
this.region.getOldestHfileTs(true);
-153} catch (IOException ioe) {
-154  LOG.error("Could not load HFile 
info ", ioe);
-155}
-156long now = 
EnvironmentEdgeManager.currentTime();
-157return now - lastMajorCompactionTs;
-158  }
-159
-160  @Override
-161  public long getNumCompactionsFailed() 
{
-162return 
this.region.compactionsFailed.get();
-163  }
-164
-165  @Override
-166  public long getMaxStoreFileAge() {
-167return maxStoreFileAge;
-168  }
-169
-170  @Override
-171  public long getMinStoreFileAge() {
-172return minStoreFileAge;
-173  }
-174
-175  @Override
-176  public long getAvgStoreFileAge() {
-177return avgStoreFileAge;
-178  }
-179
-180  @Override
-181  public long getNumReferenceFiles() {
-182return numReferenceFiles;
-183  }
-184
-185  @Override
-186  public int getRegionHashCode() {
-187return this.region.hashCode();
-188  }
-189
-190  public class 
HRegionMetricsWrapperRunnable implements Runnable {
+058  private ScheduledFuture? 
regionMetricsUpdateTask;
+059
+060  public MetricsRegionWrapperImpl(HRegion 
region) {
+061this.region = region;
+062this.executor = 
CompatibilitySingletonFactory.getInstance(MetricsExecutor.class).getExecutor();
+063this.runnable = new 
HRegionMetricsWrapperRunnable();
+064this.regionMetricsUpdateTask = 
this.executor.scheduleWithFixedDelay(this.runnable, PERIOD,
+065  PERIOD, TimeUnit.SECONDS);
+066  }
+067
+068 

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-26 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/21766f4a/devapidocs/src-html/org/apache/hadoop/hbase/client/ColumnFamilyDescriptorBuilder.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/ColumnFamilyDescriptorBuilder.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/ColumnFamilyDescriptorBuilder.html
index e1f7fd7..031b981 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/ColumnFamilyDescriptorBuilder.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/ColumnFamilyDescriptorBuilder.html
@@ -33,1335 +33,1354 @@
 025import java.util.Map;
 026import java.util.Set;
 027import java.util.function.Function;
-028import 
org.apache.hadoop.hbase.HConstants;
-029import 
org.apache.hadoop.hbase.KeepDeletedCells;
-030import 
org.apache.hadoop.hbase.MemoryCompactionPolicy;
-031import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-032import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-033import 
org.apache.hadoop.hbase.exceptions.HBaseException;
-034import 
org.apache.hadoop.hbase.io.compress.Compression;
-035import 
org.apache.hadoop.hbase.io.encoding.DataBlockEncoding;
-036import 
org.apache.hadoop.hbase.regionserver.BloomType;
-037import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-038import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema;
-039import 
org.apache.hadoop.hbase.util.Bytes;
-040import 
org.apache.hadoop.hbase.util.PrettyPrinter;
-041import 
org.apache.hadoop.hbase.util.PrettyPrinter.Unit;
-042
-043@InterfaceAudience.Public
-044public class 
ColumnFamilyDescriptorBuilder {
-045  // For future backward compatibility
-046
-047  // Version  3 was when column names 
become byte arrays and when we picked up
-048  // Time-to-live feature.  Version 4 was 
when we moved to byte arrays, HBASE-82.
-049  // Version  5 was when bloom filter 
descriptors were removed.
-050  // Version  6 adds metadata as a map 
where keys and values are byte[].
-051  // Version  7 -- add new compression 
and hfile blocksize to HColumnDescriptor (HBASE-1217)
-052  // Version  8 -- reintroduction of 
bloom filters, changed from boolean to enum
-053  // Version  9 -- add data block 
encoding
-054  // Version 10 -- change metadata to 
standard type.
-055  // Version 11 -- add column family 
level configuration.
-056  private static final byte 
COLUMN_DESCRIPTOR_VERSION = (byte) 11;
-057
-058  @InterfaceAudience.Private
-059  public static final String 
IN_MEMORY_COMPACTION = "IN_MEMORY_COMPACTION";
-060  private static final Bytes 
IN_MEMORY_COMPACTION_BYTES = new Bytes(Bytes.toBytes(IN_MEMORY_COMPACTION));
-061
-062  @InterfaceAudience.Private
-063  public static final String IN_MEMORY = 
HConstants.IN_MEMORY;
-064  private static final Bytes 
IN_MEMORY_BYTES = new Bytes(Bytes.toBytes(IN_MEMORY));
-065
-066  // These constants are used as FileInfo 
keys
-067  @InterfaceAudience.Private
-068  public static final String COMPRESSION 
= "COMPRESSION";
-069  private static final Bytes 
COMPRESSION_BYTES = new Bytes(Bytes.toBytes(COMPRESSION));
-070  @InterfaceAudience.Private
-071  public static final String 
COMPRESSION_COMPACT = "COMPRESSION_COMPACT";
-072  private static final Bytes 
COMPRESSION_COMPACT_BYTES = new Bytes(Bytes.toBytes(COMPRESSION_COMPACT));
-073  @InterfaceAudience.Private
-074  public static final String 
DATA_BLOCK_ENCODING = "DATA_BLOCK_ENCODING";
-075  private static final Bytes 
DATA_BLOCK_ENCODING_BYTES = new Bytes(Bytes.toBytes(DATA_BLOCK_ENCODING));
-076  /**
-077   * Key for the BLOCKCACHE attribute. A 
more exact name would be
-078   * CACHE_DATA_ON_READ because this flag 
sets whether or not we cache DATA
-079   * blocks. We always cache INDEX and 
BLOOM blocks; caching these blocks cannot
-080   * be disabled.
-081   */
-082  @InterfaceAudience.Private
-083  public static final String BLOCKCACHE = 
"BLOCKCACHE";
-084  private static final Bytes 
BLOCKCACHE_BYTES = new Bytes(Bytes.toBytes(BLOCKCACHE));
-085  @InterfaceAudience.Private
-086  public static final String 
CACHE_DATA_ON_WRITE = "CACHE_DATA_ON_WRITE";
-087  private static final Bytes 
CACHE_DATA_ON_WRITE_BYTES = new Bytes(Bytes.toBytes(CACHE_DATA_ON_WRITE));
-088  @InterfaceAudience.Private
-089  public static final String 
CACHE_INDEX_ON_WRITE = "CACHE_INDEX_ON_WRITE";
-090  private static final Bytes 
CACHE_INDEX_ON_WRITE_BYTES = new Bytes(Bytes.toBytes(CACHE_INDEX_ON_WRITE));
-091  @InterfaceAudience.Private
-092  public static final String 
CACHE_BLOOMS_ON_WRITE = "CACHE_BLOOMS_ON_WRITE";
-093  private static final Bytes 
CACHE_BLOOMS_ON_WRITE_BYTES = new 
Bytes(Bytes.toBytes(CACHE_BLOOMS_ON_WRITE));
-094  @InterfaceAudience.Private
-095  public static final String 
EVICT_BLOCKS_ON_CLOSE = "EVICT_BLOCKS_ON_CLOSE";
-096  private static final Bytes 
EVICT_BLOCKS_ON_CLOSE_BYTES = new 
Bytes(Bytes.toBytes(EVICT_BLOCKS_ON_CLOSE));
-097  /**
-098 

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-24 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2d5075d7/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html
index afd9ccc..904b921 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html
@@ -30,1916 +30,1984 @@
 022import 
org.apache.hadoop.hbase.shaded.com.google.common.base.Preconditions;
 023
 024import java.io.IOException;
-025import java.util.ArrayList;
-026import java.util.Arrays;
-027import java.util.HashSet;
-028import java.util.Iterator;
-029import java.util.List;
-030import java.util.Map;
-031import java.util.Set;
-032import 
java.util.concurrent.atomic.AtomicBoolean;
-033import 
java.util.concurrent.atomic.AtomicInteger;
-034import 
java.util.concurrent.atomic.AtomicLong;
-035import java.util.stream.Collectors;
-036import java.util.stream.Stream;
-037import 
java.util.concurrent.ConcurrentHashMap;
-038import 
java.util.concurrent.CopyOnWriteArrayList;
-039import java.util.concurrent.DelayQueue;
-040import java.util.concurrent.TimeUnit;
-041
-042import org.apache.commons.logging.Log;
-043import 
org.apache.commons.logging.LogFactory;
-044import 
org.apache.hadoop.conf.Configuration;
-045import 
org.apache.hadoop.hbase.HConstants;
-046import 
org.apache.hadoop.hbase.ProcedureInfo;
-047import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-048import 
org.apache.hadoop.hbase.classification.InterfaceStability;
-049import 
org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException;
-050import 
org.apache.hadoop.hbase.procedure2.Procedure.LockState;
-051import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore;
-052import 
org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator;
-053import 
org.apache.hadoop.hbase.procedure2.util.DelayedUtil;
-054import 
org.apache.hadoop.hbase.procedure2.util.DelayedUtil.DelayedWithTimeout;
-055import 
org.apache.hadoop.hbase.procedure2.util.StringUtils;
-056import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState;
-057import 
org.apache.hadoop.hbase.security.User;
-058import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-059import 
org.apache.hadoop.hbase.util.NonceKey;
-060import 
org.apache.hadoop.hbase.util.Pair;
-061import 
org.apache.hadoop.hbase.util.Threads;
-062
-063/**
-064 * Thread Pool that executes the 
submitted procedures.
-065 * The executor has a ProcedureStore 
associated.
-066 * Each operation is logged and on 
restart the pending procedures are resumed.
-067 *
-068 * Unless the Procedure code throws an 
error (e.g. invalid user input)
-069 * the procedure will complete (at some 
point in time), On restart the pending
-070 * procedures are resumed and the once 
failed will be rolledback.
-071 *
-072 * The user can add procedures to the 
executor via submitProcedure(proc)
-073 * check for the finished state via 
isFinished(procId)
-074 * and get the result via 
getResult(procId)
-075 */
-076@InterfaceAudience.Private
-077@InterfaceStability.Evolving
-078public class 
ProcedureExecutorTEnvironment {
-079  private static final Log LOG = 
LogFactory.getLog(ProcedureExecutor.class);
-080
-081  public static final String 
CHECK_OWNER_SET_CONF_KEY = "hbase.procedure.check.owner.set";
-082  private static final boolean 
DEFAULT_CHECK_OWNER_SET = false;
-083
-084  public static final String 
WORKER_KEEP_ALIVE_TIME_CONF_KEY =
-085  
"hbase.procedure.worker.keep.alive.time.msec";
-086  private static final long 
DEFAULT_WORKER_KEEP_ALIVE_TIME = Long.MAX_VALUE;
-087
-088  Testing testing = null;
-089  public static class Testing {
-090protected boolean killIfSuspended = 
false;
-091protected boolean 
killBeforeStoreUpdate = false;
-092protected boolean 
toggleKillBeforeStoreUpdate = false;
-093
-094protected boolean 
shouldKillBeforeStoreUpdate() {
-095  final boolean kill = 
this.killBeforeStoreUpdate;
-096  if 
(this.toggleKillBeforeStoreUpdate) {
-097this.killBeforeStoreUpdate = 
!kill;
-098LOG.warn("Toggle KILL before 
store update to: " + this.killBeforeStoreUpdate);
-099  }
-100  return kill;
-101}
-102
-103protected boolean 
shouldKillBeforeStoreUpdate(final boolean isSuspended) {
-104  return (isSuspended  
!killIfSuspended) ? false : shouldKillBeforeStoreUpdate();
-105}
-106  }
-107
-108  public interface 
ProcedureExecutorListener {
-109void procedureLoaded(long procId);
-110void procedureAdded(long procId);
-111void procedureFinished(long 
procId);
-112  }
-113
-114  /**

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-23 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0383a9c2/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.html
index 15fec0a..d30e78b 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.html
@@ -630,684 +630,653 @@
 622}
 623  }
 624
-625  /**
-626   * Sends an MERGE REGIONS RPC to the 
specified server to merge the specified
-627   * regions.
-628   * p
-629   * A region server could reject the 
close request because it either does not
-630   * have the specified region.
-631   * @param server server to merge 
regions
-632   * @param region_a region to merge
-633   * @param region_b region to merge
-634   * @param forcible true if do a 
compulsory merge, otherwise we will only merge
-635   *  two adjacent regions
-636   * @throws IOException
-637   */
-638  public void sendRegionsMerge(ServerName 
server, HRegionInfo region_a,
-639  HRegionInfo region_b, boolean 
forcible, final User user) throws IOException {
-640if (server == null)
-641  throw new 
NullPointerException("Passed server is null");
-642if (region_a == null || region_b == 
null)
-643  throw new 
NullPointerException("Passed region is null");
-644AdminService.BlockingInterface admin 
= getRsAdmin(server);
-645if (admin == null) {
-646  throw new IOException("Attempting 
to send MERGE REGIONS RPC to server "
-647  + server.toString() + " for 
region "
-648  + 
region_a.getRegionNameAsString() + ","
-649  + 
region_b.getRegionNameAsString()
-650  + " failed because no RPC 
connection found to this server");
-651}
-652HBaseRpcController controller = 
newRpcController();
-653ProtobufUtil.mergeRegions(controller, 
admin, region_a, region_b, forcible, user);
-654  }
-655
-656  @VisibleForTesting
-657  public void 
moveFromOnlineToDeadServers(final ServerName sn) {
-658synchronized (onlineServers) {
-659  if 
(!this.onlineServers.containsKey(sn)) {
-660LOG.warn("Expiration of " + sn + 
" but server not online");
-661  }
-662  // Remove the server from the known 
servers lists and update load info BUT
-663  // add to deadservers first; do 
this so it'll show in dead servers list if
-664  // not in online servers list.
-665  this.deadservers.add(sn);
-666  this.onlineServers.remove(sn);
-667  onlineServers.notifyAll();
-668}
-669this.rsAdmins.remove(sn);
-670  }
-671
-672  public synchronized void 
processDeadServer(final ServerName serverName, boolean shouldSplitWal) {
-673// When assignment manager is 
cleaning up the zookeeper nodes and rebuilding the
-674// in-memory region states, region 
servers could be down. Meta table can and
-675// should be re-assigned, log 
splitting can be done too. However, it is better to
-676// wait till the cleanup is done 
before re-assigning user regions.
-677//
-678// We should not wait in the server 
shutdown handler thread since it can clog
-679// the handler threads and meta table 
could not be re-assigned in case
-680// the corresponding server is down. 
So we queue them up here instead.
-681if 
(!master.getAssignmentManager().isFailoverCleanupDone()) {
-682  requeuedDeadServers.put(serverName, 
shouldSplitWal);
-683  return;
-684}
-685
-686this.deadservers.add(serverName);
-687
master.getAssignmentManager().submitServerCrash(serverName, shouldSplitWal);
-688  }
-689
-690  /**
-691   * Process the servers which died 
during master's initialization. It will be
-692   * called after HMaster#assignMeta and 
AssignmentManager#joinCluster.
-693   * */
-694  synchronized void 
processQueuedDeadServers() {
-695if 
(!master.isServerCrashProcessingEnabled()) {
-696  LOG.info("Master hasn't enabled 
ServerShutdownHandler");
-697}
-698IteratorServerName 
serverIterator = queuedDeadServers.iterator();
-699while (serverIterator.hasNext()) {
-700  ServerName tmpServerName = 
serverIterator.next();
-701  expireServer(tmpServerName);
-702  serverIterator.remove();
-703  
requeuedDeadServers.remove(tmpServerName);
-704}
-705
-706if 
(!master.getAssignmentManager().isFailoverCleanupDone()) {
-707  LOG.info("AssignmentManager hasn't 
finished failover cleanup; waiting");
-708}
-709
-710for (Map.EntryServerName, 
Boolean entry : requeuedDeadServers.entrySet()) {
-711  processDeadServer(entry.getKey(), 
entry.getValue());
-712}
-713requeuedDeadServers.clear();
-714  }
-715
-716  /*
-717   * Remove the server from the drain 
list.
-718   */
-719  public boolean 
removeServerFromDrainList(final ServerName sn) {
-720// Warn 

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-22 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f391bcef/devapidocs/org/apache/hadoop/hbase/client/RegionServerCallable.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/RegionServerCallable.html 
b/devapidocs/org/apache/hadoop/hbase/client/RegionServerCallable.html
index 91b6520..f04789a 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/RegionServerCallable.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/RegionServerCallable.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":6,"i12":10,"i13":10,"i14":10,"i15":6,"i16":10,"i17":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":6,"i13":10,"i14":10,"i15":10,"i16":6,"i17":10,"i18":10};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],4:["t3","Abstract Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -123,7 +123,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public abstract class RegionServerCallableT,S
+public abstract class RegionServerCallableT,S
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements RetryingCallableT
 Implementations make a RPC call against a RegionService via 
a protobuf Service.
@@ -165,21 +165,25 @@ implements 
+private int
+priority
+
+
 private byte[]
 row
 
-
+
 protected 
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController
 rpcController
 This is 99% of the time a HBaseRpcController but also used 
doing Coprocessor Endpoints and in
  this case, it is a ServerRpcControllable which is not a 
HBaseRpcController.
 
 
-
+
 protected S
 stub
 
-
+
 private TableName
 tableName
 
@@ -203,6 +207,13 @@ implements 
+RegionServerCallable(Connectionconnection,
+TableNametableName,
+byte[]row,
+
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllerrpcController,
+intpriority)
+
 
 
 
@@ -241,63 +252,67 @@ implements getLocation()
 
 
+protected int
+getPriority()
+
+
 byte[]
 getRow()
 
-
+
 protected 
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController
 getRpcController()
 
-
+
 protected CellScanner
 getRpcControllerCellScanner()
 Get the RpcController CellScanner.
 
 
-
+
 protected S
 getStub()
 
-
+
 TableName
 getTableName()
 
-
+
 void
 prepare(booleanreload)
 Prepare by setting up any connections to servers, etc., 
ahead of call invocation.
 
 
-
+
 protected abstract T
 rpcCall()
 Run the RPC call.
 
 
-
+
 protected void
 setLocation(HRegionLocationlocation)
 
-
+
 protected void
 setRpcControllerCellScanner(CellScannercellScanner)
 
-
+
 protected void
 setStub(Sstub)
 
-
+
 protected abstract void
 setStubByServiceName(ServerNameserviceName)
 Set the RCP client stub
 
 
-
+
 long
 sleep(longpause,
  inttries)
 
-
+
 void
 throwable(http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
 title="class or interface in java.lang">Throwablet,
  booleanretrying)
@@ -333,7 +348,7 @@ implements 
 
 connection
-private finalConnection connection
+private finalConnection connection
 
 
 
@@ -342,7 +357,7 @@ implements 
 
 tableName
-private finalTableName tableName
+private finalTableName tableName
 
 
 
@@ -351,7 +366,7 @@ implements 
 
 row
-private finalbyte[] row
+private finalbyte[] row
 
 
 
@@ -360,7 +375,7 @@ implements 
 
 location
-protectedHRegionLocation location
+protectedHRegionLocation location
 Some subclasses want to set their own location. Make it 
protected.
 
 
@@ -370,21 +385,30 @@ implements 
 
 stub
-protectedS stub
+protectedS stub
 
 
 
 
 
-
+
 
 rpcController
-protected 
finalorg.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController rpcController
+protected 
finalorg.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController rpcController
 This is 99% of the time a HBaseRpcController but also used 
doing Coprocessor Endpoints and in
  this case, it is a ServerRpcControllable which is not a HBaseRpcController.
  Can be null!
 
 
+
+
+
+
+
+priority
+privateint priority
+
+
 
 
 
@@ -396,10 +420,10 @@ implements 
 
 
-
+
 
 RegionServerCallable
-publicRegionServerCallable(Connectionconnection,
+publicRegionServerCallable(Connectionconnection,
 TableNametableName,
 byte[]row,
 
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllerrpcController)
@@ -412,6 +436,19 @@ implements 
+
+
+
+
+RegionServerCallable
+publicRegionServerCallable(Connectionconnection,
+TableNametableName,
+byte[]row,
+

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-21 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ca5b0275/devapidocs/org/apache/hadoop/hbase/security/visibility/VisibilityReplicationEndpoint.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/security/visibility/VisibilityReplicationEndpoint.html
 
b/devapidocs/org/apache/hadoop/hbase/security/visibility/VisibilityReplicationEndpoint.html
index 20d91ce..07607fc 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/security/visibility/VisibilityReplicationEndpoint.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/security/visibility/VisibilityReplicationEndpoint.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -109,12 +109,12 @@ var activeTableTab = "activeTableTab";
 
 
 All Implemented Interfaces:
-com.google.common.util.concurrent.Service, ReplicationEndpoint, ReplicationPeerConfigListener
+ReplicationEndpoint, ReplicationPeerConfigListener, 
org.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.Service
 
 
 
 @InterfaceAudience.Private
-public class VisibilityReplicationEndpoint
+public class VisibilityReplicationEndpoint
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements ReplicationEndpoint
 
@@ -137,11 +137,11 @@ implements ReplicationEndpoint.Context, ReplicationEndpoint.ReplicateContext
 
 
-
+
 
 
-Nested classes/interfaces inherited from 
interfacecom.google.common.util.concurrent.Service
-com.google.common.util.concurrent.Service.State
+Nested classes/interfaces inherited from 
interfaceorg.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.Service
+org.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.Service.Listener,
 
org.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.Service.State
 
 
 
@@ -203,65 +203,84 @@ implements Method and Description
 
 
+void
+addListener(org.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.Service.Listenerlistener,
+   http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/Executor.html?is-external=true;
 title="class or interface in 
java.util.concurrent">Executorexecutor)
+
+
+void
+awaitRunning()
+
+
+void
+awaitRunning(longl,
+http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true;
 title="class or interface in 
java.util.concurrent">TimeUnittimeUnit)
+
+
+void
+awaitTerminated()
+
+
+void
+awaitTerminated(longl,
+   http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true;
 title="class or interface in 
java.util.concurrent">TimeUnittimeUnit)
+
+
 boolean
 canReplicateToSameCluster()
 Whether or not, the replication endpoint can replicate to 
it's source cluster with the same
  UUID
 
 
-
+
+http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
 title="class or interface in java.lang">Throwable
+failureCause()
+
+
 http://docs.oracle.com/javase/8/docs/api/java/util/UUID.html?is-external=true;
 title="class or interface in java.util">UUID
 getPeerUUID()
 Returns a UUID of the provided peer id.
 
 
-
+
 WALEntryFilter
 getWALEntryfilter()
 Returns a WALEntryFilter to use for filtering out 
WALEntries from the log.
 
 
-
+
 void
 init(ReplicationEndpoint.Contextcontext)
 Initialize the replication endpoint with the given 
context.
 
 
-
+
 boolean
 isRunning()
 
-
+
 void
 peerConfigUpdated(ReplicationPeerConfigrpc)
 Callback method for when users update the 
ReplicationPeerConfig for this peer
 
 
-
+
 boolean
 replicate(ReplicationEndpoint.ReplicateContextreplicateContext)
 Replicate the given set of entries (in the context) to the 
other cluster.
 
 
-
-com.google.common.util.concurrent.ListenableFuturecom.google.common.util.concurrent.Service.State
-start()
+
+org.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.Service
+startAsync()
 
-
-com.google.common.util.concurrent.Service.State
-startAndWait()
-
-
-com.google.common.util.concurrent.Service.State
+
+org.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.Service.State
 state()
 
-
-com.google.common.util.concurrent.ListenableFuturecom.google.common.util.concurrent.Service.State
-stop()
-
-
-com.google.common.util.concurrent.Service.State
-stopAndWait()
+
+org.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.Service
+stopAsync()
 
 
 
@@ -291,7 +310,7 @@ implements 
 
 LOG
-private static finalorg.apache.commons.logging.Log LOG
+private static 

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-19 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/9eba7fcf/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmplImpl.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmplImpl.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmplImpl.html
index 00312b7..962b1ef 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmplImpl.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmplImpl.html
@@ -189,318 +189,318 @@
 181AssignmentManager assignmentManager = 
master.getAssignmentManager();
 182
 183// 96, 1
-184jamonWriter.write("!--[if 
IE]\n!DOCTYPE html\n![endif]--\n?xml version=\"1.0\" 
encoding=\"UTF-8\" ?\nhtml lang=\"en\"\n  head\n
meta charset=\"utf-8\"\ntitle");
-185// 103, 12
+184jamonWriter.write("!DOCTYPE 
html\n?xml version=\"1.0\" encoding=\"UTF-8\" ?\nhtml 
lang=\"en\"\n  head\nmeta charset=\"utf-8\"\n
title");
+185// 101, 12
 186if (master.isActiveMaster() )
 187{
-188  // 103, 43
+188  // 101, 43
 189  jamonWriter.write("Master: ");
 190}
-191// 103, 51
+191// 101, 51
 192else
 193{
-194  // 103, 58
+194  // 101, 58
 195  jamonWriter.write("Backup Master: 
");
 196}
-197// 103, 79
+197// 101, 79
 198jamonWriter.write("\n");
-199// 104, 5
+199// 102, 5
 200
org.jamon.escaping.Escaping.NONE.write(org.jamon.emit.StandardEmitter.valueOf(master.getServerName().getHostname()),
 jamonWriter);
-201// 104, 47
+201// 102, 47
 202jamonWriter.write("/title\n   
 meta name=\"viewport\" content=\"width=device-width, 
initial-scale=1.0\"\nmeta name=\"description\" content=\"\"\n   
 link href=\"/static/css/bootstrap.min.css\" rel=\"stylesheet\"\n
link href=\"/static/css/bootstrap-theme.min.css\" rel=\"stylesheet\"\n  
  link href=\"/static/css/hbase.css\" rel=\"stylesheet\"\n  
/head\n\n  body\n\ndiv class=\"navbar  navbar-fixed-top 
navbar-default\"\ndiv class=\"container-fluid\"\n   
 div class=\"navbar-header\"\nbutton 
type=\"button\" class=\"navbar-toggle\" data-toggle=\"collapse\" 
data-target=\".navbar-collapse\"\nspan 
class=\"icon-bar\"/span\nspan 
class=\"icon-bar\"/span\nspan 
class=\"icon-bar\"/span\n/button
 t;\na class=\"navbar-brand\" 
href=\"/master-status\"img src=\"/static/hbase_logo_small.png\" 
alt=\"HBase Logo\"//a\n/div\n
div class=\"collapse navbar-collapse\"\nul 
class=\"nav navbar-nav\"\nli class=\"active\"a 
href=\"/\"Home/a/li\nlia 
href=\"/tablesDetailed.jsp\"Table Details/a/li\n
");
-203// 128, 17
+203// 126, 17
 204if (master.isActiveMaster() )
 205{
-206  // 128, 48
+206  // 126, 48
 207  jamonWriter.write("\n   
 lia href=\"/procedures.jsp\"Procedures amp; 
Locks/a/li\n");
 208}
-209// 130, 23
+209// 128, 23
 210jamonWriter.write("\n
lia href=\"/processMaster.jsp\"Process 
Metrics/a/li\nlia 
href=\"/logs/\"Local Logs/a/li\n
lia href=\"/logLevel\"Log Level/a/li\n  
  lia href=\"/dump\"Debug Dump/a/li\n   
 lia href=\"/jmx\"Metrics Dump/a/li\n   
 ");
-211// 136, 17
+211// 134, 17
 212if 
(HBaseConfiguration.isShowConfInServlet())
 213{
-214  // 136, 64
+214  // 134, 64
 215  jamonWriter.write("\n   
 lia href=\"/conf\"HBase Configuration/a/li\n   
 ");
 216}
-217// 138, 23
+217// 136, 23
 218jamonWriter.write("\n   
/ul\n/div!--/.nav-collapse --\n
/div\n/div\n\ndiv class=\"container-fluid 
content\"\n\t");
-219// 145, 2
+219// 143, 2
 220if (master.isActiveMaster() )
 221{
-222  // 145, 33
+222  // 143, 33
 223  jamonWriter.write("\n
div class=\"row inner_header\"\ndiv 
class=\"page-header\"\nh1Master small");
-224  // 148, 35
+224  // 146, 35
 225  
org.jamon.escaping.Escaping.NONE.write(org.jamon.emit.StandardEmitter.valueOf(master.getServerName().getHostname()),
 jamonWriter);
-226  // 148, 77
+226  // 146, 77
 227  
jamonWriter.write("/small/h1\n/div\n
/div\n\ndiv class=\"row\"\n!-- Various 
warnings that cluster admins should be aware of --\n");
-228  // 154, 9
+228  // 152, 9
 229  if (JvmVersion.isBadJvmVersion() 
)
 230  {
-231// 154, 45
+231// 152, 45
 232jamonWriter.write("\n  
div class=\"alert alert-error\"\n  Your current JVM version 
");
-233// 156, 36

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-15 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/17128d27/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
index feb42ea..4bd98f4 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
@@ -185,4189 +185,4266 @@
 177import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest;
 178import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest;
 179import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse;
-180import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest;
-181import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest;
-182import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse;
-183import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest;
-184import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos;
-185import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos;
-186import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse;
-187import 
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos;
-188import 
org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils;
-189import 
org.apache.hadoop.hbase.snapshot.HBaseSnapshotException;
-190import 
org.apache.hadoop.hbase.snapshot.RestoreSnapshotException;
-191import 
org.apache.hadoop.hbase.snapshot.SnapshotCreationException;
-192import 
org.apache.hadoop.hbase.snapshot.UnknownSnapshotException;
-193import 
org.apache.hadoop.hbase.util.Addressing;
-194import 
org.apache.hadoop.hbase.util.Bytes;
-195import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-196import 
org.apache.hadoop.hbase.util.ForeignExceptionUtil;
-197import 
org.apache.hadoop.hbase.util.Pair;
-198import 
org.apache.hadoop.hbase.zookeeper.MasterAddressTracker;
-199import 
org.apache.hadoop.hbase.zookeeper.MetaTableLocator;
-200import 
org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher;
-201import 
org.apache.hadoop.ipc.RemoteException;
-202import 
org.apache.hadoop.util.StringUtils;
-203import 
org.apache.zookeeper.KeeperException;
-204
-205import 
com.google.common.annotations.VisibleForTesting;
-206import com.google.protobuf.Descriptors;
-207import com.google.protobuf.Message;
-208import 
com.google.protobuf.RpcController;
-209import java.util.stream.Collectors;
-210
-211/**
-212 * HBaseAdmin is no longer a client API. 
It is marked InterfaceAudience.Private indicating that
-213 * this is an HBase-internal class as 
defined in
-214 * 
https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/InterfaceClassification.html
-215 * There are no guarantees for backwards 
source / binary compatibility and methods or class can
-216 * change or go away without 
deprecation.
-217 * Use {@link Connection#getAdmin()} to 
obtain an instance of {@link Admin} instead of constructing
-218 * an HBaseAdmin directly.
-219 *
-220 * pConnection should be an 
iunmanaged/i connection obtained via
-221 * {@link 
ConnectionFactory#createConnection(Configuration)}
-222 *
-223 * @see ConnectionFactory
-224 * @see Connection
-225 * @see Admin
-226 */
-227@InterfaceAudience.Private
-228@InterfaceStability.Evolving
-229public class HBaseAdmin implements Admin 
{
-230  private static final Log LOG = 
LogFactory.getLog(HBaseAdmin.class);
-231
-232  private static final String 
ZK_IDENTIFIER_PREFIX =  "hbase-admin-on-";
+180import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest;
+181import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse;
+182import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest;
+183import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest;
+184import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse;
+185import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest;
+186import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos;
+187import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos;
+188import 
org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse;
+189import 
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos;
+190import 
org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils;
+191import 

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-12 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2777c693/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestStore.MyList.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestStore.MyList.html 
b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestStore.MyList.html
index 526f5da..3b51f87 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestStore.MyList.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestStore.MyList.html
@@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab";
 
 
 PrevClass
-NextClass
+NextClass
 
 
 Frames
@@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static class TestStore.MyListT
+private static class TestStore.MyListT
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Object
 implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListT
 
@@ -143,7 +143,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html
 delegatee
 
 
-private http://docs.oracle.com/javase/8/docs/api/java/util/function/Consumer.html?is-external=true;
 title="class or interface in java.util.function">Consumerhttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
 title="class or interface in java.lang">Integer
+private TestStore.MyListHook
 hookAtAdd
 
 
@@ -161,7 +161,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html
 Constructor and Description
 
 
-MyList(http://docs.oracle.com/javase/8/docs/api/java/util/function/Consumer.html?is-external=true;
 title="class or interface in java.util.function">Consumerhttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
 title="class or interface in 
java.lang">IntegerhookAtAdd)
+MyList(TestStore.MyListHookhookAtAdd)
 
 
 
@@ -323,7 +323,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html
 
 
 delegatee
-private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListT delegatee
+private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
 title="class or interface in java.util">ListT delegatee
 
 
 
@@ -332,7 +332,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html
 
 
 hookAtAdd
-private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/function/Consumer.html?is-external=true;
 title="class or interface in java.util.function">Consumerhttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
 title="class or interface in java.lang">Integer hookAtAdd
+private finalTestStore.MyListHook hookAtAdd
 
 
 
@@ -343,13 +343,13 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html
 
 
 Constructor Detail
-
+
 
 
 
 
 MyList
-MyList(http://docs.oracle.com/javase/8/docs/api/java/util/function/Consumer.html?is-external=true;
 title="class or interface in java.util.function">Consumerhttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
 title="class or interface in java.lang">IntegerhookAtAdd)
+MyList(TestStore.MyListHookhookAtAdd)
 
 
 
@@ -366,7 +366,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html
 
 
 size
-publicintsize()
+publicintsize()
 
 Specified by:
 http://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true#size--;
 title="class or interface in java.util">sizein 
interfacehttp://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
 title="class or interface in java.util">CollectionT
@@ -381,7 +381,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html
 
 
 isEmpty
-publicbooleanisEmpty()
+publicbooleanisEmpty()
 
 Specified by:
 http://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true#isEmpty--;
 title="class or interface in java.util">isEmptyin 
interfacehttp://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
 title="class or interface in java.util">CollectionT
@@ -396,7 +396,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html
 
 
 contains
-publicbooleancontains(http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Objecto)
+publicbooleancontains(http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
 title="class or interface in java.lang">Objecto)
 
 Specified by:
 http://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true#contains-java.lang.Object-;
 title="class or interface in java.util">containsin 
interfacehttp://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
 title="class or interface in 

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-11 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/90c7dfe4/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html
index 75db22d..99a09f9 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html
@@ -37,2710 +37,2816 @@
 029import java.util.List;
 030import java.util.Map;
 031import java.util.Optional;
-032import 
java.util.concurrent.CompletableFuture;
-033import java.util.concurrent.TimeUnit;
-034import 
java.util.concurrent.atomic.AtomicReference;
-035import java.util.function.BiConsumer;
-036import java.util.regex.Pattern;
-037import java.util.stream.Collectors;
-038
-039import 
com.google.common.annotations.VisibleForTesting;
-040
-041import io.netty.util.Timeout;
-042import io.netty.util.TimerTask;
-043
-044import java.util.stream.Stream;
-045
-046import org.apache.commons.io.IOUtils;
-047import org.apache.commons.logging.Log;
-048import 
org.apache.commons.logging.LogFactory;
-049import 
org.apache.hadoop.hbase.ClusterStatus;
-050import 
org.apache.hadoop.hbase.HRegionInfo;
-051import 
org.apache.hadoop.hbase.HRegionLocation;
-052import 
org.apache.hadoop.hbase.MetaTableAccessor;
-053import 
org.apache.hadoop.hbase.MetaTableAccessor.QueryType;
-054import 
org.apache.hadoop.hbase.NotServingRegionException;
-055import 
org.apache.hadoop.hbase.ProcedureInfo;
-056import 
org.apache.hadoop.hbase.RegionLoad;
-057import 
org.apache.hadoop.hbase.RegionLocations;
-058import 
org.apache.hadoop.hbase.ServerName;
-059import 
org.apache.hadoop.hbase.NamespaceDescriptor;
-060import 
org.apache.hadoop.hbase.HConstants;
-061import 
org.apache.hadoop.hbase.TableExistsException;
-062import 
org.apache.hadoop.hbase.TableName;
-063import 
org.apache.hadoop.hbase.AsyncMetaTableAccessor;
-064import 
org.apache.hadoop.hbase.TableNotDisabledException;
-065import 
org.apache.hadoop.hbase.TableNotEnabledException;
-066import 
org.apache.hadoop.hbase.TableNotFoundException;
-067import 
org.apache.hadoop.hbase.UnknownRegionException;
-068import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-069import 
org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder;
-070import 
org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder;
-071import 
org.apache.hadoop.hbase.client.Scan.ReadType;
-072import 
org.apache.hadoop.hbase.client.replication.ReplicationSerDeHelper;
-073import 
org.apache.hadoop.hbase.client.replication.TableCFs;
-074import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-075import 
org.apache.hadoop.hbase.ipc.HBaseRpcController;
-076import 
org.apache.hadoop.hbase.quotas.QuotaFilter;
-077import 
org.apache.hadoop.hbase.quotas.QuotaSettings;
-078import 
org.apache.hadoop.hbase.quotas.QuotaTableUtil;
-079import 
org.apache.hadoop.hbase.replication.ReplicationException;
-080import 
org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
-081import 
org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
-082import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback;
-083import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-084import 
org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
-085import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
-086import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest;
-087import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse;
-088import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest;
-089import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse;
-090import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest;
-091import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionResponse;
-092import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionRequest;
-093import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionResponse;
-094import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest;
-095import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse;
-096import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequest;
-097import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse;
-098import 

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0821e51a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html
index 71844ce..75db22d 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html
@@ -105,2564 +105,2642 @@
 097import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse;
 098import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionRequest;
 099import 
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionResponse;
-100import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription;
-101import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema;
-102import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType;
-103import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest;
-104import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse;
-105import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest;
-106import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse;
-107import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest;
-108import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse;
-109import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest;
-110import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse;
-111import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest;
-112import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse;
-113import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest;
-114import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse;
-115import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest;
-116import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse;
-117import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest;
-118import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse;
-119import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest;
-120import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse;
-121import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest;
-122import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse;
-123import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest;
-124import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse;
-125import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest;
-126import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse;
-127import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest;
-128import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse;
-129import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest;
-130import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse;
-131import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest;
-132import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse;
-133import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest;
-134import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse;
-135import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest;
-136import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse;
-137import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest;
-138import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse;
-139import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest;
-140import 

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2d27954a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html
index 16c0042..71844ce 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html
@@ -126,2499 +126,2543 @@
 118import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse;
 119import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest;
 120import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse;
-121import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest;
-122import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse;
-123import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest;
-124import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse;
-125import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest;
-126import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse;
-127import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest;
-128import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse;
-129import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest;
-130import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse;
-131import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest;
-132import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse;
-133import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest;
-134import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse;
-135import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest;
-136import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse;
-137import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest;
-138import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse;
-139import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest;
-140import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse;
-141import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest;
-142import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse;
-143import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest;
-144import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse;
-145import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest;
-146import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse;
-147import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest;
-148import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse;
-149import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest;
-150import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse;
-151import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest;
-152import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse;
-153import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest;
-154import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse;
-155import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest;
-156import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse;
-157import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest;
-158import 
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse;
-159import 

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-08 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/9fb0764b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html 
b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
index 369cb85..021cf88 100644
--- a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
+++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
@@ -2986,10 +2986,25 @@ service.
 
 
 
+(package private) static HTableDescriptor
+HBaseAdmin.getHTableDescriptor(TableNametableName,
+   Connectionconnection,
+   RpcRetryingCallerFactoryrpcCallerFactory,
+   RpcControllerFactoryrpcControllerFactory,
+   intoperationTimeout,
+   intrpcTimeout)
+Deprecated.
+since 2.0 version and will 
be removed in 3.0 version.
+ use HBaseAdmin.getTableDescriptor(TableName,
+ Connection, 
RpcRetryingCallerFactory,RpcControllerFactory,int,int)
+
+
+
+
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
 title="class or interface in java.util">Optionalhttp://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
 title="class or interface in java.lang">Long
 AsyncHBaseAdmin.getLastMajorCompactionTimestamp(TableNametableName)
 
-
+
 long
 Admin.getLastMajorCompactionTimestamp(TableNametableName)
 Get the timestamp of the last major compaction for the 
passed table
@@ -2998,74 +3013,74 @@ service.
  or 0 if no such HFile could be found.
 
 
-
+
 long
 HBaseAdmin.getLastMajorCompactionTimestamp(TableNametableName)
 
-
+
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
 title="class or interface in java.util">Optionalhttp://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
 title="class or interface in java.lang">Long
 AsyncAdmin.getLastMajorCompactionTimestamp(TableNametableName)
 Get the timestamp of the last major compaction for the 
passed table.
 
 
-
+
 http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
 title="class or interface in java.util">Optionalhttp://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
 title="class or interface in java.lang">Long
 RawAsyncHBaseAdmin.getLastMajorCompactionTimestamp(TableNametableName)
 
-
+
 private http://docs.oracle.com/javase/8/docs/api/java/util/function/Consumer.html?is-external=true;
 title="class or interface in java.util.function">Consumerhttp://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
 title="class or interface in java.lang">Long
 AsyncProcess.getLogger(TableNametableName,
  longmax)
 
-
+
 private HRegionInfo
 HBaseAdmin.getMobRegionInfo(TableNametableName)
 
-
+
 int
 MetaCache.getNumberOfCachedRegionLocations(TableNametableName)
 Return the number of cached region for a table.
 
 
-
+
 (package private) int
 ConnectionImplementation.getNumberOfCachedRegionLocations(TableNametableName)
 
-
+
 default RawAsyncTable
 AsyncConnection.getRawTable(TableNametableName)
 Retrieve an RawAsyncTable implementation 
for accessing a table.
 
 
-
+
 AsyncTableBuilderRawAsyncTable
 AsyncConnection.getRawTableBuilder(TableNametableName)
 Returns an AsyncTableBuilder for creating 
RawAsyncTable.
 
 
-
+
 AsyncTableBuilderRawAsyncTable
 AsyncConnectionImpl.getRawTableBuilder(TableNametableName)
 
-
+
 http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Mapbyte[],RegionLoad
 Admin.getRegionLoad(ServerNamesn,
  TableNametableName)
 Get RegionLoad of all regions hosted on a 
regionserver for a table.
 
 
-
+
 http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
 title="class or interface in java.util">Mapbyte[],RegionLoad
 HBaseAdmin.getRegionLoad(ServerNamesn,
  TableNametableName)
 
-
+
 HRegionLocation
 ConnectionImplementation.getRegionLocation(TableNametableName,
  byte[]row,
  booleanreload)
 
-
+
 HRegionLocation
 ClusterConnection.getRegionLocation(TableNametableName,
  byte[]row,
@@ -3073,26 +3088,26 @@ service.
 Find region location hosting passed row
 
 
-
+
 (package private) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
 

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-07 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b3b50f22/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/bucket/FileIOEngine.FileWriteAccessor.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/bucket/FileIOEngine.FileWriteAccessor.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/bucket/FileIOEngine.FileWriteAccessor.html
index ba2c5f6..803999d 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/bucket/FileIOEngine.FileWriteAccessor.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/bucket/FileIOEngine.FileWriteAccessor.html
@@ -60,209 +60,222 @@
 052  private FileReadAccessor readAccessor = 
new FileReadAccessor();
 053  private FileWriteAccessor writeAccessor 
= new FileWriteAccessor();
 054
-055  public FileIOEngine(long capacity, 
String... filePaths) throws IOException {
-056this.sizePerFile = capacity / 
filePaths.length;
-057this.capacity = this.sizePerFile * 
filePaths.length;
-058this.filePaths = filePaths;
-059this.fileChannels = new 
FileChannel[filePaths.length];
-060this.rafs = new 
RandomAccessFile[filePaths.length];
-061for (int i = 0; i  
filePaths.length; i++) {
-062  String filePath = filePaths[i];
-063  try {
-064rafs[i] = new 
RandomAccessFile(filePath, "rw");
-065long totalSpace = new 
File(filePath).getTotalSpace();
-066if (totalSpace  sizePerFile) 
{
-067  // The next setting length will 
throw exception,logging this message
-068  // is just used for the detail 
reason of exception,
-069  String msg = "Only " + 
StringUtils.byteDesc(totalSpace)
-070  + " total space under " + 
filePath + ", not enough for requested "
-071  + 
StringUtils.byteDesc(sizePerFile);
-072  LOG.warn(msg);
-073}
-074rafs[i].setLength(sizePerFile);
-075fileChannels[i] = 
rafs[i].getChannel();
-076LOG.info("Allocating cache " + 
StringUtils.byteDesc(sizePerFile)
-077+ ", on the path:" + 
filePath);
-078  } catch (IOException fex) {
-079LOG.error("Failed allocating 
cache on " + filePath, fex);
-080shutdown();
-081throw fex;
-082  }
-083}
-084  }
-085
-086  @Override
-087  public String toString() {
-088return "ioengine=" + 
this.getClass().getSimpleName() + ", paths="
-089+ Arrays.asList(filePaths) + ", 
capacity=" + String.format("%,d", this.capacity);
-090  }
-091
-092  /**
-093   * File IO engine is always able to 
support persistent storage for the cache
-094   * @return true
-095   */
-096  @Override
-097  public boolean isPersistent() {
-098return true;
-099  }
-100
-101  /**
-102   * Transfers data from file to the 
given byte buffer
-103   * @param offset The offset in the file 
where the first byte to be read
-104   * @param length The length of buffer 
that should be allocated for reading
-105   *   from the file 
channel
-106   * @return number of bytes read
-107   * @throws IOException
+055  public FileIOEngine(long capacity, 
boolean maintainPersistence, String... filePaths)
+056  throws IOException {
+057this.sizePerFile = capacity / 
filePaths.length;
+058this.capacity = this.sizePerFile * 
filePaths.length;
+059this.filePaths = filePaths;
+060this.fileChannels = new 
FileChannel[filePaths.length];
+061if (!maintainPersistence) {
+062  for (String filePath : filePaths) 
{
+063File file = new File(filePath);
+064if (file.exists()) {
+065  if (LOG.isDebugEnabled()) {
+066LOG.debug("File " + filePath 
+ " already exists. Deleting!!");
+067  }
+068  file.delete();
+069  // If deletion fails still we 
can manage with the writes
+070}
+071  }
+072}
+073this.rafs = new 
RandomAccessFile[filePaths.length];
+074for (int i = 0; i  
filePaths.length; i++) {
+075  String filePath = filePaths[i];
+076  try {
+077rafs[i] = new 
RandomAccessFile(filePath, "rw");
+078long totalSpace = new 
File(filePath).getTotalSpace();
+079if (totalSpace  sizePerFile) 
{
+080  // The next setting length will 
throw exception,logging this message
+081  // is just used for the detail 
reason of exception,
+082  String msg = "Only " + 
StringUtils.byteDesc(totalSpace)
+083  + " total space under " + 
filePath + ", not enough for requested "
+084  + 
StringUtils.byteDesc(sizePerFile);
+085  LOG.warn(msg);
+086}
+087rafs[i].setLength(sizePerFile);
+088fileChannels[i] = 
rafs[i].getChannel();
+089LOG.info("Allocating cache " + 
StringUtils.byteDesc(sizePerFile)
+090+ ", on the path:" + 
filePath);
+091  } catch (IOException fex) {
+092LOG.error("Failed allocating 
cache on " + filePath, fex);
+093shutdown();

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-07-05 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ca9f6925/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTable.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTable.html 
b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTable.html
index 1fd28df..ec7f66d 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTable.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTable.html
@@ -49,7 +49,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-PrevClass
+PrevClass
 NextClass
 
 
@@ -393,7 +393,7 @@ extends 
 
-PrevClass
+PrevClass
 NextClass
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ca9f6925/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncAdmin.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncAdmin.html 
b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncAdmin.html
index 3b69c91..d3bf407 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncAdmin.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncAdmin.html
@@ -98,6 +98,27 @@
 
 Uses of AsyncAdmin in org.apache.hadoop.hbase.client
 
+Classes in org.apache.hadoop.hbase.client
 with type parameters of type AsyncAdmin
+
+Modifier and Type
+Interface and Description
+
+
+
+interface
+AsyncAdminBuilderT 
extends AsyncAdmin
+For creating AsyncAdmin.
+
+
+
+(package private) class
+AsyncAdminBuilderBaseT 
extends AsyncAdmin
+Base class for all asynchronous admin builders.
+
+
+
+
+
 Classes in org.apache.hadoop.hbase.client
 that implement AsyncAdmin
 
 Modifier and Type
@@ -110,6 +131,12 @@
 The implementation of AsyncAdmin.
 
 
+
+class
+RawAsyncHBaseAdmin
+The implementation of AsyncAdmin.
+
+
 
 
 
@@ -121,7 +148,7 @@
 
 
 protected AsyncAdmin
-AsyncHBaseAdmin.ProcedureBiConsumer.admin
+RawAsyncHBaseAdmin.ProcedureBiConsumer.admin
 
 
 
@@ -133,14 +160,16 @@
 
 
 
-AsyncAdmin
+default AsyncAdmin
 AsyncConnection.getAdmin()
-Retrieve an AsyncAdmin implementation to administer an 
HBase cluster.
+Retrieve an AsyncAdmin 
implementation to administer an HBase cluster.
 
 
 
-AsyncAdmin
-AsyncConnectionImpl.getAdmin()
+default AsyncAdmin
+AsyncConnection.getAdmin(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true;
 title="class or interface in 
java.util.concurrent">ExecutorServicepool)
+Retrieve an AsyncAdmin 
implementation to administer an HBase cluster.
+
 
 
 
@@ -151,62 +180,62 @@
 
 
 
-AddColumnFamilyProcedureBiConsumer(AsyncAdminadmin,
+AddColumnFamilyProcedureBiConsumer(AsyncAdminadmin,
   TableNametableName)
 
 
-CreateNamespaceProcedureBiConsumer(AsyncAdminadmin,
+CreateNamespaceProcedureBiConsumer(AsyncAdminadmin,
   http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">StringnamespaceName)
 
 
-CreateTableProcedureBiConsumer(AsyncAdminadmin,
+CreateTableProcedureBiConsumer(AsyncAdminadmin,
   TableNametableName)
 
 
-DeleteColumnFamilyProcedureBiConsumer(AsyncAdminadmin,
+DeleteColumnFamilyProcedureBiConsumer(AsyncAdminadmin,
  TableNametableName)
 
 
-DeleteNamespaceProcedureBiConsumer(AsyncAdminadmin,
+DeleteNamespaceProcedureBiConsumer(AsyncAdminadmin,
   http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">StringnamespaceName)
 
 
-DeleteTableProcedureBiConsumer(AsyncAdminadmin,
+DeleteTableProcedureBiConsumer(AsyncAdminadmin,
   TableNametableName)
 
 
-DisableTableProcedureBiConsumer(AsyncAdminadmin,
+DisableTableProcedureBiConsumer(AsyncAdminadmin,
TableNametableName)
 
 
-EnableTableProcedureBiConsumer(AsyncAdminadmin,
+EnableTableProcedureBiConsumer(AsyncAdminadmin,
   TableNametableName)
 
 
-MergeTableRegionProcedureBiConsumer(AsyncAdminadmin,
+MergeTableRegionProcedureBiConsumer(AsyncAdminadmin,
TableNametableName)
 
 
-ModifyColumnFamilyProcedureBiConsumer(AsyncAdminadmin,
+ModifyColumnFamilyProcedureBiConsumer(AsyncAdminadmin,
  TableNametableName)
 
 
-ModifyNamespaceProcedureBiConsumer(AsyncAdminadmin,
+ModifyNamespaceProcedureBiConsumer(AsyncAdminadmin,
   http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in 
java.lang">StringnamespaceName)
 
 
-NamespaceProcedureBiConsumer(AsyncAdminadmin,
+NamespaceProcedureBiConsumer(AsyncAdminadmin,
 http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface 

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-06-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8e3b63ca/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html
index c895448..545d4da 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html
@@ -1294,425 +1294,426 @@
 1286  }
 1287
 1288  // We normalize locality to be a 
score between 0 and 1.0 representing how good it
-1289  // is compared to how good it 
could be
-1290  locality /= bestLocality;
-1291}
-1292
-1293@Override
-1294protected void regionMoved(int 
region, int oldServer, int newServer) {
-1295  int oldEntity = type == 
LocalityType.SERVER ? oldServer : cluster.serverIndexToRackIndex[oldServer];
-1296  int newEntity = type == 
LocalityType.SERVER ? newServer : cluster.serverIndexToRackIndex[newServer];
-1297  if (this.services == null) {
-1298return;
-1299  }
-1300  double localityDelta = 
getWeightedLocality(region, newEntity) - getWeightedLocality(region, 
oldEntity);
-1301  double normalizedDelta = 
localityDelta / bestLocality;
-1302  locality += normalizedDelta;
-1303}
-1304
-1305@Override
-1306double cost() {
-1307  return 1 - locality;
-1308}
-1309
-1310private int 
getMostLocalEntityForRegion(int region) {
-1311  return 
cluster.getOrComputeRegionsToMostLocalEntities(type)[region];
-1312}
-1313
-1314private double 
getWeightedLocality(int region, int entity) {
-1315  return 
cluster.getOrComputeWeightedLocality(region, entity, type);
-1316}
-1317
-1318  }
-1319
-1320  static class 
ServerLocalityCostFunction extends LocalityBasedCostFunction {
-1321
-1322private static final String 
LOCALITY_COST_KEY = "hbase.master.balancer.stochastic.localityCost";
-1323private static final float 
DEFAULT_LOCALITY_COST = 25;
-1324
-1325
ServerLocalityCostFunction(Configuration conf, MasterServices srv) {
-1326  super(
-1327  conf,
-1328  srv,
-1329  LocalityType.SERVER,
-1330  LOCALITY_COST_KEY,
-1331  DEFAULT_LOCALITY_COST
-1332  );
-1333}
-1334
-1335@Override
-1336int regionIndexToEntityIndex(int 
region) {
-1337  return 
cluster.regionIndexToServerIndex[region];
-1338}
-1339  }
-1340
-1341  static class RackLocalityCostFunction 
extends LocalityBasedCostFunction {
-1342
-1343private static final String 
RACK_LOCALITY_COST_KEY = "hbase.master.balancer.stochastic.rackLocalityCost";
-1344private static final float 
DEFAULT_RACK_LOCALITY_COST = 15;
-1345
-1346public 
RackLocalityCostFunction(Configuration conf, MasterServices services) {
-1347  super(
-1348  conf,
-1349  services,
-1350  LocalityType.RACK,
-1351  RACK_LOCALITY_COST_KEY,
-1352  DEFAULT_RACK_LOCALITY_COST
-1353  );
-1354}
-1355
-1356@Override
-1357int regionIndexToEntityIndex(int 
region) {
-1358  return 
cluster.getRackForRegion(region);
-1359}
-1360  }
-1361
-1362  /**
-1363   * Base class the allows writing costs 
functions from rolling average of some
-1364   * number from RegionLoad.
-1365   */
-1366  abstract static class 
CostFromRegionLoadFunction extends CostFunction {
-1367
-1368private ClusterStatus clusterStatus 
= null;
-1369private MapString, 
DequeBalancerRegionLoad loads = null;
-1370private double[] stats = null;
-1371
CostFromRegionLoadFunction(Configuration conf) {
-1372  super(conf);
-1373}
-1374
-1375void setClusterStatus(ClusterStatus 
status) {
-1376  this.clusterStatus = status;
-1377}
-1378
-1379void setLoads(MapString, 
DequeBalancerRegionLoad l) {
-1380  this.loads = l;
-1381}
-1382
-1383@Override
-1384double cost() {
-1385  if (clusterStatus == null || loads 
== null) {
-1386return 0;
-1387  }
-1388
-1389  if (stats == null || stats.length 
!= cluster.numServers) {
-1390stats = new 
double[cluster.numServers];
-1391  }
-1392
-1393  for (int i =0; i  
stats.length; i++) {
-1394//Cost this server has from 
RegionLoad
-1395long cost = 0;
-1396
-1397// for every region on this 
server get the rl
-1398for(int 
regionIndex:cluster.regionsPerServer[i]) {
-1399  
CollectionBalancerRegionLoad regionLoadList =  
cluster.regionLoads[regionIndex];
-1400
-1401  // Now if we found a region 
load get the type of cost that was requested.
-1402  if (regionLoadList != null) 
{
-1403  

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-06-26 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aecb1286/devapidocs/src-html/org/apache/hadoop/hbase/Version.html
--
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/Version.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/Version.html
index 114f6d1..a811283 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/Version.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/Version.html
@@ -16,11 +16,11 @@
 008@InterfaceAudience.Private
 009public class Version {
 010  public static final String version = 
"3.0.0-SNAPSHOT";
-011  public static final String revision = 
"0e8e176ebd3bd17d969d17ce2b0aa3dafb93fa22";
+011  public static final String revision = 
"2d781aa15c9535c3845692a275c8babca2c45568";
 012  public static final String user = 
"jenkins";
-013  public static final String date = "Sun 
Jun 25 14:39:31 UTC 2017";
+013  public static final String date = "Mon 
Jun 26 14:39:40 UTC 2017";
 014  public static final String url = 
"git://asf920.gq1.ygridcore.net/home/jenkins/jenkins-slave/workspace/hbase_generate_website/hbase";
-015  public static final String srcChecksum 
= "f767e20900e65fc306ff06b652532cf8";
+015  public static final String srcChecksum 
= "06d75479e61d981c6c5cc84833e9a9c7";
 016}
 
 



[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-06-20 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a719cd00/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.PrimaryRegionCountSkewCostFunction.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.PrimaryRegionCountSkewCostFunction.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.PrimaryRegionCountSkewCostFunction.html
index 6de986f..c895448 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.PrimaryRegionCountSkewCostFunction.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.PrimaryRegionCountSkewCostFunction.html
@@ -26,1592 +26,1693 @@
 018package 
org.apache.hadoop.hbase.master.balancer;
 019
 020import java.util.ArrayDeque;
-021import java.util.Arrays;
-022import java.util.Collection;
-023import java.util.Deque;
-024import java.util.HashMap;
-025import java.util.LinkedList;
-026import java.util.List;
-027import java.util.Map;
-028import java.util.Map.Entry;
-029import java.util.Random;
-030
-031import org.apache.commons.logging.Log;
-032import 
org.apache.commons.logging.LogFactory;
-033import 
org.apache.hadoop.conf.Configuration;
-034import 
org.apache.hadoop.hbase.ClusterStatus;
-035import 
org.apache.hadoop.hbase.HBaseInterfaceAudience;
-036import 
org.apache.hadoop.hbase.HConstants;
-037import 
org.apache.hadoop.hbase.HRegionInfo;
-038import 
org.apache.hadoop.hbase.RegionLoad;
-039import 
org.apache.hadoop.hbase.ServerLoad;
-040import 
org.apache.hadoop.hbase.ServerName;
-041import 
org.apache.hadoop.hbase.TableName;
-042import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
-043import 
org.apache.hadoop.hbase.master.MasterServices;
-044import 
org.apache.hadoop.hbase.master.RegionPlan;
-045import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action;
-046import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action.Type;
-047import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.AssignRegionAction;
-048import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.MoveRegionAction;
-049import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.SwapRegionsAction;
-050import 
org.apache.hadoop.hbase.util.Bytes;
-051import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-052
-053import com.google.common.collect.Lists;
-054
-055/**
-056 * pThis is a best effort load 
balancer. Given a Cost function F(C) =gt; x It will
-057 * randomly try and mutate the cluster to 
Cprime. If F(Cprime) lt; F(C) then the
-058 * new cluster state becomes the plan. It 
includes costs functions to compute the cost of:/p
-059 * ul
-060 * liRegion Load/li
-061 * liTable Load/li
-062 * liData Locality/li
-063 * liMemstore Sizes/li
-064 * liStorefile Sizes/li
-065 * /ul
-066 *
-067 *
-068 * pEvery cost function returns a 
number between 0 and 1 inclusive; where 0 is the lowest cost
-069 * best solution, and 1 is the highest 
possible cost and the worst solution.  The computed costs are
-070 * scaled by their respective 
multipliers:/p
+021import java.util.ArrayList;
+022import java.util.Arrays;
+023import java.util.Collection;
+024import java.util.Collections;
+025import java.util.Deque;
+026import java.util.HashMap;
+027import java.util.LinkedList;
+028import java.util.List;
+029import java.util.Map;
+030import java.util.Map.Entry;
+031import java.util.Random;
+032
+033import org.apache.commons.logging.Log;
+034import 
org.apache.commons.logging.LogFactory;
+035import 
org.apache.hadoop.conf.Configuration;
+036import 
org.apache.hadoop.hbase.ClusterStatus;
+037import 
org.apache.hadoop.hbase.HBaseInterfaceAudience;
+038import 
org.apache.hadoop.hbase.HConstants;
+039import 
org.apache.hadoop.hbase.HRegionInfo;
+040import 
org.apache.hadoop.hbase.RegionLoad;
+041import 
org.apache.hadoop.hbase.ServerLoad;
+042import 
org.apache.hadoop.hbase.ServerName;
+043import 
org.apache.hadoop.hbase.TableName;
+044import 
org.apache.hadoop.hbase.classification.InterfaceAudience;
+045import 
org.apache.hadoop.hbase.master.MasterServices;
+046import 
org.apache.hadoop.hbase.master.RegionPlan;
+047import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action;
+048import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action.Type;
+049import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.AssignRegionAction;
+050import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.LocalityType;
+051import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.MoveRegionAction;
+052import 
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.SwapRegionsAction;
+053import 
org.apache.hadoop.hbase.util.Bytes;
+054import 
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
+055

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-06-10 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/476c54ed/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaTableUtil.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaTableUtil.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaTableUtil.html
index 509b93c..3c6f9b8 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaTableUtil.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaTableUtil.html
@@ -53,606 +53,717 @@
 045import 
org.apache.hadoop.hbase.client.ResultScanner;
 046import 
org.apache.hadoop.hbase.client.Scan;
 047import 
org.apache.hadoop.hbase.client.Table;
-048import 
org.apache.hadoop.hbase.filter.CompareFilter;
-049import 
org.apache.hadoop.hbase.filter.Filter;
-050import 
org.apache.hadoop.hbase.filter.FilterList;
-051import 
org.apache.hadoop.hbase.filter.QualifierFilter;
-052import 
org.apache.hadoop.hbase.filter.RegexStringComparator;
-053import 
org.apache.hadoop.hbase.filter.RowFilter;
-054import 
org.apache.hadoop.hbase.protobuf.ProtobufMagic;
-055import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString;
-056import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
-057import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations;
-058import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-059import 
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
-060import 
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos;
-061import 
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesResponse;
-062import 
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesResponse;
-063import 
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaSnapshotsResponse;
-064import 
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaSnapshotsResponse.TableQuotaSnapshot;
-065import 
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesResponse.RegionSizes;
-066import 
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas;
-067import 
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceQuota;
-068import 
org.apache.hadoop.hbase.util.Bytes;
-069import 
org.apache.hadoop.hbase.util.Strings;
-070
-071/**
-072 * Helper class to interact with the 
quota table.
-073 * pre
-074 * ROW-KEY  FAM/QUAL
DATA
-075 *   n.lt;namespacegt; q:s  
   lt;global-quotasgt;
-076 *   t.lt;namespacegt; u:p  
  lt;namespace-quota policygt;
-077 *   t.lt;tablegt; q:s  
   lt;global-quotasgt;
-078 *   t.lt;tablegt; u:p  
  lt;table-quota policygt;
-079 *   u.lt;usergt;  q:s  
   lt;global-quotasgt;
-080 *   u.lt;usergt;  
q:s.lt;tablegt; lt;table-quotasgt;
-081 *   u.lt;usergt;  
q:s.lt;nsgt;:   lt;namespace-quotasgt;
-082 * /pre
-083 */
-084@InterfaceAudience.Private
-085@InterfaceStability.Evolving
-086public class QuotaTableUtil {
-087  private static final Log LOG = 
LogFactory.getLog(QuotaTableUtil.class);
-088
-089  /** System table for quotas */
-090  public static final TableName 
QUOTA_TABLE_NAME =
-091  
TableName.valueOf(NamespaceDescriptor.SYSTEM_NAMESPACE_NAME_STR, "quota");
-092
-093  protected static final byte[] 
QUOTA_FAMILY_INFO = Bytes.toBytes("q");
-094  protected static final byte[] 
QUOTA_FAMILY_USAGE = Bytes.toBytes("u");
-095  protected static final byte[] 
QUOTA_QUALIFIER_SETTINGS = Bytes.toBytes("s");
-096  protected static final byte[] 
QUOTA_QUALIFIER_SETTINGS_PREFIX = Bytes.toBytes("s.");
-097  protected static final byte[] 
QUOTA_QUALIFIER_POLICY = Bytes.toBytes("p");
-098  protected static final String 
QUOTA_POLICY_COLUMN =
-099  Bytes.toString(QUOTA_FAMILY_USAGE) 
+ ":" + Bytes.toString(QUOTA_QUALIFIER_POLICY);
-100  protected static final byte[] 
QUOTA_USER_ROW_KEY_PREFIX = Bytes.toBytes("u.");
-101  protected static final byte[] 
QUOTA_TABLE_ROW_KEY_PREFIX = Bytes.toBytes("t.");
-102  protected static final byte[] 
QUOTA_NAMESPACE_ROW_KEY_PREFIX = Bytes.toBytes("n.");
-103
-104  /* 
=
-105   *  Quota "settings" helpers
-106   */
-107  public static Quotas 
getTableQuota(final Connection connection, final TableName table)
-108  throws IOException {
-109return getQuotas(connection, 
getTableRowKey(table));
-110  }
-111
-112  public static Quotas 
getNamespaceQuota(final Connection connection, final String namespace)
-113  throws IOException {
-114return getQuotas(connection, 
getNamespaceRowKey(namespace));
-115  }
-116
-117  public static Quotas getUserQuota(final 
Connection connection, final String user)
-118  throws IOException {
-119return getQuotas(connection, 
getUserRowKey(user));
-120  }
-121
-122  public static Quotas getUserQuota(final 

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-06-09 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/77a552c4/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationState.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationState.html 
b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationState.html
index a969f92..a37ea12 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationState.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationState.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static enum HBaseAdmin.ReplicationState
+private static enum HBaseAdmin.ReplicationState
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
 title="class or interface in java.lang">EnumHBaseAdmin.ReplicationState
 This enum indicates the current state of the replication 
for a given table.
 
@@ -214,7 +214,7 @@ the order they are declared.
 
 
 ENABLED
-public static finalHBaseAdmin.ReplicationState 
ENABLED
+public static finalHBaseAdmin.ReplicationState 
ENABLED
 
 
 
@@ -223,7 +223,7 @@ the order they are declared.
 
 
 MIXED
-public static finalHBaseAdmin.ReplicationState 
MIXED
+public static finalHBaseAdmin.ReplicationState 
MIXED
 
 
 
@@ -232,7 +232,7 @@ the order they are declared.
 
 
 DISABLED
-public static finalHBaseAdmin.ReplicationState 
DISABLED
+public static finalHBaseAdmin.ReplicationState 
DISABLED
 
 
 
@@ -249,7 +249,7 @@ the order they are declared.
 
 
 values
-public staticHBaseAdmin.ReplicationState[]values()
+public staticHBaseAdmin.ReplicationState[]values()
 Returns an array containing the constants of this enum 
type, in
 the order they are declared.  This method may be used to iterate
 over the constants as follows:
@@ -269,7 +269,7 @@ for (HBaseAdmin.ReplicationState c : 
HBaseAdmin.ReplicationState.values())
 
 
 valueOf
-public staticHBaseAdmin.ReplicationStatevalueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringname)
+public staticHBaseAdmin.ReplicationStatevalueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">Stringname)
 Returns the enum constant of this type with the specified 
name.
 The string must match exactly an identifier used to declare an
 enum constant in this type.  (Extraneous whitespace characters are 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/77a552c4/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html
 
b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html
index 1417d3a..9bfcda0 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html
@@ -127,7 +127,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-private static class HBaseAdmin.RestoreSnapshotFuture
+private static class HBaseAdmin.RestoreSnapshotFuture
 extends HBaseAdmin.TableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void
 
 
@@ -240,7 +240,7 @@ extends 
 
 RestoreSnapshotFuture
-publicRestoreSnapshotFuture(HBaseAdminadmin,
+publicRestoreSnapshotFuture(HBaseAdminadmin,
  
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionsnapshot,
  TableNametableName,
  
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponseresponse)
@@ -252,7 +252,7 @@ extends 
 
 RestoreSnapshotFuture
-publicRestoreSnapshotFuture(HBaseAdminadmin,
+publicRestoreSnapshotFuture(HBaseAdminadmin,
  TableNametableName,
  http://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
 title="class or interface in java.lang">LongprocId)
 
@@ -271,7 +271,7 @@ extends 
 
 getOperationType
-publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringgetOperationType()
+publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
 title="class or interface in java.lang">StringgetOperationType()
 
 Specified by:
 getOperationTypein
 classHBaseAdmin.TableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
 title="class or interface in java.lang">Void

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/77a552c4/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.TableFuture.TableWaitForStateCallable.html

[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.

2017-06-07 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b44796ef/devapidocs/org/apache/hadoop/hbase/backup/class-use/BackupRestoreConstants.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/backup/class-use/BackupRestoreConstants.html
 
b/devapidocs/org/apache/hadoop/hbase/backup/class-use/BackupRestoreConstants.html
index c810478..32b1ef9 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/backup/class-use/BackupRestoreConstants.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/backup/class-use/BackupRestoreConstants.html
@@ -4,7 +4,7 @@
 
 
 
-Uses of Interface org.apache.hadoop.hbase.backup.BackupRestoreConstants 
(Apache HBase 2.0.0-SNAPSHOT API)
+Uses of Interface org.apache.hadoop.hbase.backup.BackupRestoreConstants 
(Apache HBase 3.0.0-SNAPSHOT API)
 
 
 
@@ -12,7 +12,7 @@
 
 
-Uses of Class org.apache.hadoop.hbase.backup.BackupRestoreFactory 
(Apache HBase 2.0.0-SNAPSHOT API)
+Uses of Class org.apache.hadoop.hbase.backup.BackupRestoreFactory 
(Apache HBase 3.0.0-SNAPSHOT API)