[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/346adc37/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html index cbc2ae9..9f4d04d 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html @@ -692,20 +692,20 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage +org.apache.hadoop.hbase.regionserver.MemStoreCompactor.Action +org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection org.apache.hadoop.hbase.regionserver.FlushType -org.apache.hadoop.hbase.regionserver.ScannerContext.NextState -org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status -org.apache.hadoop.hbase.regionserver.StoreScanner.StoreScannerCompactionRace -org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType -org.apache.hadoop.hbase.regionserver.BloomType org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope org.apache.hadoop.hbase.regionserver.Region.Operation +org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType org.apache.hadoop.hbase.regionserver.ScanType -org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection -org.apache.hadoop.hbase.regionserver.Region.FlushResult.Result -org.apache.hadoop.hbase.regionserver.MemStoreCompactor.Action org.apache.hadoop.hbase.regionserver.RegionOpeningState +org.apache.hadoop.hbase.regionserver.BloomType +org.apache.hadoop.hbase.regionserver.ScannerContext.NextState +org.apache.hadoop.hbase.regionserver.StoreScanner.StoreScannerCompactionRace +org.apache.hadoop.hbase.regionserver.Region.FlushResult.Result +org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status +org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage http://git-wip-us.apache.org/repos/asf/hbase-site/blob/346adc37/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/NormalUserScanQueryMatcher.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/NormalUserScanQueryMatcher.html b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/NormalUserScanQueryMatcher.html index e8361b7..f1461a6 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/NormalUserScanQueryMatcher.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/NormalUserScanQueryMatcher.html @@ -380,7 +380,7 @@ extends Specified by: beforeShippedin interfaceShipperListener Overrides: -beforeShippedin classScanQueryMatcher +beforeShippedin classUserScanQueryMatcher Throws: http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException http://git-wip-us.apache.org/repos/asf/hbase-site/blob/346adc37/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/RawScanQueryMatcher.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/RawScanQueryMatcher.html b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/RawScanQueryMatcher.html index cb4af05..307e792 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/RawScanQueryMatcher.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/RawScanQueryMatcher.html @@ -242,14 +242,14 @@ extends UserScanQueryMatcher -create, getFilter, getNextKeyHint, hasNullColumnInQuery, isUserScan, matchColumn, moreRowsMayExistAfter, moreRowsMayExistsAfter +beforeShipped, create, getFilter, getNextKeyHint, hasNullColumnInQuery, isUserScan, matchColumn, moreRowsMayExistAfter, moreRowsMayExistsAfter Methods inherited from classorg.apache.hadoop.hbase.regionserver.querymatcher.ScanQueryMatcher -beforeShipped, checkColumn, checkDeleted, clearCurrentRow, compareKeyForNextColumn, compareKeyForNextRow, createStartKeyFromRow, currentRow, getKeyForNextColumn, getStartKey, getTrackers, preCheck, setToNewRow +checkColumn, checkDeleted, clearCurrentRow, compareKeyForNextColumn, compareK eyForNextRow, createStartKeyFromRow, currentRow, getKeyForNextColumn, getStartKey, getTrackers, preCheck,
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2aec596e/devapidocs/org/apache/hadoop/hbase/client/Admin.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/Admin.html b/devapidocs/org/apache/hadoop/hbase/client/Admin.html index 281158f..acc88a2 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/Admin.html +++ b/devapidocs/org/apache/hadoop/hbase/client/Admin.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":6,"i1":6,"i2":6,"i3":38,"i4":6,"i5":6,"i6":18,"i7":18,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6,"i20":6,"i21":6,"i22":6,"i23":6,"i24":6,"i25":6,"i26":6,"i27":6,"i28":6,"i29":6,"i30":6,"i31":6,"i32":50,"i33":50,"i34":50,"i35":6,"i36":6,"i37":6,"i38":50,"i39":6,"i40":38,"i41":6,"i42":6,"i43":6,"i44":6,"i45":6,"i46":6,"i47":6,"i48":6,"i49":6,"i50":6,"i51":38,"i52":38,"i53":6,"i54":6,"i55":18,"i56":6,"i57":6,"i58":6,"i59":38,"i60":38,"i61":6,"i62":6,"i63":18,"i64":6,"i65":6,"i66":6,"i67":38,"i68":38,"i69":6,"i70":6,"i71":6,"i72":6,"i73":38,"i74":6,"i75":6,"i76":6,"i77":6,"i78":6,"i79":6,"i80":6,"i81":6,"i82":6,"i83":6,"i84":6,"i85":6,"i86":6,"i87":6,"i88":6,"i89":6,"i90":6,"i91":18,"i92":6,"i93":38,"i94":38,"i95":38,"i96":6,"i97":6,"i98":6,"i99":6,"i100":6,"i101":6,"i102":6,"i103":6,"i104":6,"i105":6,"i106":6,"i107":6,"i108":6,"i109":6,"i110":6,"i111":6,"i112":6,"i113":6,"i114":6,"i115":18,"i116":18,"i117":18,"i118" :6,"i119":6,"i120":6,"i121":6,"i122":6,"i123":6,"i124":6,"i125":6,"i126":6,"i127":6,"i128":6,"i129":38,"i130":6,"i131":6,"i132":6,"i133":6,"i134":6,"i135":6,"i136":38,"i137":38,"i138":38,"i139":38,"i140":38,"i141":6,"i142":6,"i143":6,"i144":6,"i145":6,"i146":6,"i147":6,"i148":6,"i149":38,"i150":6,"i151":6,"i152":38,"i153":6,"i154":6,"i155":6,"i156":6,"i157":6,"i158":38,"i159":6,"i160":38,"i161":6,"i162":6,"i163":6,"i164":6,"i165":18,"i166":18,"i167":6,"i168":6,"i169":6,"i170":6,"i171":6,"i172":6,"i173":6,"i174":6,"i175":6,"i176":6,"i177":6,"i178":6,"i179":6,"i180":6,"i181":6,"i182":6,"i183":6,"i184":6,"i185":6,"i186":6,"i187":6,"i188":38,"i189":38,"i190":6,"i191":6,"i192":6,"i193":6,"i194":6,"i195":6,"i196":6,"i197":6,"i198":6,"i199":6,"i200":18}; +var methods = {"i0":6,"i1":6,"i2":6,"i3":38,"i4":6,"i5":6,"i6":18,"i7":18,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":38,"i18":38,"i19":38,"i20":38,"i21":6,"i22":6,"i23":6,"i24":6,"i25":6,"i26":6,"i27":6,"i28":6,"i29":6,"i30":6,"i31":6,"i32":50,"i33":50,"i34":50,"i35":6,"i36":6,"i37":6,"i38":50,"i39":6,"i40":38,"i41":6,"i42":6,"i43":6,"i44":6,"i45":6,"i46":6,"i47":6,"i48":6,"i49":6,"i50":6,"i51":38,"i52":38,"i53":6,"i54":6,"i55":18,"i56":6,"i57":6,"i58":6,"i59":38,"i60":38,"i61":6,"i62":6,"i63":18,"i64":6,"i65":6,"i66":6,"i67":38,"i68":38,"i69":6,"i70":6,"i71":6,"i72":6,"i73":38,"i74":6,"i75":6,"i76":6,"i77":6,"i78":6,"i79":6,"i80":6,"i81":6,"i82":6,"i83":6,"i84":6,"i85":6,"i86":6,"i87":6,"i88":6,"i89":6,"i90":6,"i91":18,"i92":6,"i93":38,"i94":38,"i95":38,"i96":6,"i97":6,"i98":6,"i99":6,"i100":6,"i101":6,"i102":6,"i103":6,"i104":6,"i105":6,"i106":6,"i107":6,"i108":6,"i109":6,"i110":6,"i111":6,"i112":6,"i113":6,"i114":6,"i115":18,"i116":18,"i117":18,"i 118":6,"i119":6,"i120":6,"i121":6,"i122":6,"i123":6,"i124":6,"i125":6,"i126":6,"i127":6,"i128":6,"i129":38,"i130":6,"i131":6,"i132":6,"i133":6,"i134":6,"i135":6,"i136":38,"i137":38,"i138":38,"i139":38,"i140":38,"i141":6,"i142":6,"i143":6,"i144":6,"i145":6,"i146":6,"i147":6,"i148":6,"i149":38,"i150":6,"i151":6,"i152":38,"i153":6,"i154":6,"i155":6,"i156":6,"i157":6,"i158":38,"i159":6,"i160":38,"i161":6,"i162":6,"i163":6,"i164":6,"i165":18,"i166":18,"i167":6,"i168":6,"i169":6,"i170":6,"i171":6,"i172":6,"i173":6,"i174":6,"i175":6,"i176":6,"i177":6,"i178":6,"i179":6,"i180":6,"i181":6,"i182":6,"i183":6,"i184":6,"i185":6,"i186":6,"i187":6,"i188":38,"i189":38,"i190":6,"i191":6,"i192":6,"i193":6,"i194":6,"i195":6,"i196":6,"i197":6,"i198":6,"i199":6,"i200":18}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],16:["t5","Default Methods"],32:["t6","Deprecated Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -264,28 +264,36 @@ extends void closeRegion(byte[]regionname, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringserverName) -Close a region. +Deprecated. +Since 2.0. Will be removed in 3.0. Use unassign(byte[], boolean) instead. + void closeRegion(ServerNamesn, HRegionInfohri) -Close a region. +Deprecated. +Since 2.0. Will be removed in 3.0. Use unassign(byte[], boolean) instead. + void closeRegion(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringregionname,
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1837997e/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/AnnotationReadingPriorityFunction.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/AnnotationReadingPriorityFunction.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/AnnotationReadingPriorityFunction.html index 7e411c6..39d8ea2 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/AnnotationReadingPriorityFunction.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/AnnotationReadingPriorityFunction.html @@ -41,255 +41,253 @@ 033import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest; 034import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest; 035import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest; -036import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionRequest; -037import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.GetRequest; -038import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.MutateRequest; -039import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ScanRequest; -040import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier; -041import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; -042 -043import org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting; -044import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; -045import org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat; -046import org.apache.hadoop.hbase.security.User; -047 -048/** -049 * Reads special method annotations and table names to figure a priority for use by QoS facility in -050 * ipc; e.g: rpcs to hbase:meta get priority. -051 */ -052// TODO: Remove. This is doing way too much work just to figure a priority. Do as Elliott -053// suggests and just have the client specify a priority. -054 -055//The logic for figuring out high priority RPCs is as follows: -056//1. if the method is annotated with a QosPriority of QOS_HIGH, -057// that is honored -058//2. parse out the protobuf message and see if the request is for meta -059// region, and if so, treat it as a high priority RPC -060//Some optimizations for (2) are done here - -061//Clients send the argument classname as part of making the RPC. The server -062//decides whether to deserialize the proto argument message based on the -063//pre-established set of argument classes (knownArgumentClasses below). -064//This prevents the server from having to deserialize all proto argument -065//messages prematurely. -066//All the argument classes declare a 'getRegion' method that returns a -067//RegionSpecifier object. Methods can be invoked on the returned object -068//to figure out whether it is a meta region or not. -069@InterfaceAudience.Private -070public class AnnotationReadingPriorityFunction implements PriorityFunction { -071 private static final Log LOG = -072 LogFactory.getLog(AnnotationReadingPriorityFunction.class.getName()); -073 -074 /** Used to control the scan delay, currently sqrt(numNextCall * weight) */ -075 public static final String SCAN_VTIME_WEIGHT_CONF_KEY = "hbase.ipc.server.scan.vtime.weight"; -076 -077 protected final MapString, Integer annotatedQos; -078 //We need to mock the regionserver instance for some unit tests (set via -079 //setRegionServer method. -080 private RSRpcServices rpcServices; -081 @SuppressWarnings("unchecked") -082 private final Class? extends Message[] knownArgumentClasses = new Class[]{ -083 GetRegionInfoRequest.class, -084 GetStoreFileRequest.class, -085 CloseRegionRequest.class, -086 FlushRegionRequest.class, -087 SplitRegionRequest.class, -088 CompactRegionRequest.class, -089 GetRequest.class, -090 MutateRequest.class, -091 ScanRequest.class -092 }; -093 -094 // Some caches for helping performance -095 private final MapString, Class? extends Message argumentToClassMap = new HashMap(); -096 private final MapString, MapClass? extends Message, Method methodMap = new HashMap(); +036import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.GetRequest; +037import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.MutateRequest; +038import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ScanRequest; +039import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier; +040import org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.RequestHeader; +041 +042import org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting; +043import org.apache.hadoop.hbase.shaded.com.google.protobuf.Message; +044import
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a2b2dd19/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapperImpl.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapperImpl.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapperImpl.html index 80168c0..0066ffd 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapperImpl.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapperImpl.html @@ -60,205 +60,237 @@ 052 private long minStoreFileAge; 053 private long avgStoreFileAge; 054 private long numReferenceFiles; -055 -056 private ScheduledFuture? regionMetricsUpdateTask; +055 private long maxFlushQueueSize; +056 private long maxCompactionQueueSize; 057 -058 public MetricsRegionWrapperImpl(HRegion region) { -059this.region = region; -060this.executor = CompatibilitySingletonFactory.getInstance(MetricsExecutor.class).getExecutor(); -061this.runnable = new HRegionMetricsWrapperRunnable(); -062this.regionMetricsUpdateTask = this.executor.scheduleWithFixedDelay(this.runnable, PERIOD, -063 PERIOD, TimeUnit.SECONDS); -064 } -065 -066 @Override -067 public String getTableName() { -068TableDescriptor tableDesc = this.region.getTableDescriptor(); -069if (tableDesc == null) { -070 return UNKNOWN; -071} -072return tableDesc.getTableName().getQualifierAsString(); -073 } -074 -075 @Override -076 public String getNamespace() { -077TableDescriptor tableDesc = this.region.getTableDescriptor(); -078if (tableDesc == null) { -079 return UNKNOWN; -080} -081return tableDesc.getTableName().getNamespaceAsString(); -082 } -083 -084 -085 @Override -086 public String getRegionName() { -087HRegionInfo regionInfo = this.region.getRegionInfo(); -088if (regionInfo == null) { -089 return UNKNOWN; -090} -091return regionInfo.getEncodedName(); -092 } -093 -094 @Override -095 public long getNumStores() { -096Mapbyte[],Store stores = this.region.stores; -097if (stores == null) { -098 return 0; -099} -100return stores.size(); -101 } -102 -103 @Override -104 public long getNumStoreFiles() { -105return numStoreFiles; -106 } -107 -108 @Override -109 public long getMemstoreSize() { -110return memstoreSize; -111 } -112 -113 @Override -114 public long getStoreFileSize() { -115return storeFileSize; -116 } -117 -118 @Override -119 public long getReadRequestCount() { -120return this.region.getReadRequestsCount(); -121 } -122 -123 @Override -124 public long getFilteredReadRequestCount() { -125return this.region.getFilteredReadRequestsCount(); -126 } -127 -128 @Override -129 public long getWriteRequestCount() { -130return this.region.getWriteRequestsCount(); -131 } -132 -133 @Override -134 public long getNumFilesCompacted() { -135return this.region.compactionNumFilesCompacted.get(); -136 } -137 -138 @Override -139 public long getNumBytesCompacted() { -140return this.region.compactionNumBytesCompacted.get(); -141 } -142 -143 @Override -144 public long getNumCompactionsCompleted() { -145return this.region.compactionsFinished.get(); -146 } -147 -148 @Override -149 public long getLastMajorCompactionAge() { -150long lastMajorCompactionTs = 0L; -151try { -152 lastMajorCompactionTs = this.region.getOldestHfileTs(true); -153} catch (IOException ioe) { -154 LOG.error("Could not load HFile info ", ioe); -155} -156long now = EnvironmentEdgeManager.currentTime(); -157return now - lastMajorCompactionTs; -158 } -159 -160 @Override -161 public long getNumCompactionsFailed() { -162return this.region.compactionsFailed.get(); -163 } -164 -165 @Override -166 public long getMaxStoreFileAge() { -167return maxStoreFileAge; -168 } -169 -170 @Override -171 public long getMinStoreFileAge() { -172return minStoreFileAge; -173 } -174 -175 @Override -176 public long getAvgStoreFileAge() { -177return avgStoreFileAge; -178 } -179 -180 @Override -181 public long getNumReferenceFiles() { -182return numReferenceFiles; -183 } -184 -185 @Override -186 public int getRegionHashCode() { -187return this.region.hashCode(); -188 } -189 -190 public class HRegionMetricsWrapperRunnable implements Runnable { +058 private ScheduledFuture? regionMetricsUpdateTask; +059 +060 public MetricsRegionWrapperImpl(HRegion region) { +061this.region = region; +062this.executor = CompatibilitySingletonFactory.getInstance(MetricsExecutor.class).getExecutor(); +063this.runnable = new HRegionMetricsWrapperRunnable(); +064this.regionMetricsUpdateTask = this.executor.scheduleWithFixedDelay(this.runnable, PERIOD, +065 PERIOD, TimeUnit.SECONDS); +066 } +067 +068
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/21766f4a/devapidocs/src-html/org/apache/hadoop/hbase/client/ColumnFamilyDescriptorBuilder.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/ColumnFamilyDescriptorBuilder.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/ColumnFamilyDescriptorBuilder.html index e1f7fd7..031b981 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/ColumnFamilyDescriptorBuilder.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/ColumnFamilyDescriptorBuilder.html @@ -33,1335 +33,1354 @@ 025import java.util.Map; 026import java.util.Set; 027import java.util.function.Function; -028import org.apache.hadoop.hbase.HConstants; -029import org.apache.hadoop.hbase.KeepDeletedCells; -030import org.apache.hadoop.hbase.MemoryCompactionPolicy; -031import org.apache.hadoop.hbase.classification.InterfaceAudience; -032import org.apache.hadoop.hbase.exceptions.DeserializationException; -033import org.apache.hadoop.hbase.exceptions.HBaseException; -034import org.apache.hadoop.hbase.io.compress.Compression; -035import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -036import org.apache.hadoop.hbase.regionserver.BloomType; -037import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -038import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema; -039import org.apache.hadoop.hbase.util.Bytes; -040import org.apache.hadoop.hbase.util.PrettyPrinter; -041import org.apache.hadoop.hbase.util.PrettyPrinter.Unit; -042 -043@InterfaceAudience.Public -044public class ColumnFamilyDescriptorBuilder { -045 // For future backward compatibility -046 -047 // Version 3 was when column names become byte arrays and when we picked up -048 // Time-to-live feature. Version 4 was when we moved to byte arrays, HBASE-82. -049 // Version 5 was when bloom filter descriptors were removed. -050 // Version 6 adds metadata as a map where keys and values are byte[]. -051 // Version 7 -- add new compression and hfile blocksize to HColumnDescriptor (HBASE-1217) -052 // Version 8 -- reintroduction of bloom filters, changed from boolean to enum -053 // Version 9 -- add data block encoding -054 // Version 10 -- change metadata to standard type. -055 // Version 11 -- add column family level configuration. -056 private static final byte COLUMN_DESCRIPTOR_VERSION = (byte) 11; -057 -058 @InterfaceAudience.Private -059 public static final String IN_MEMORY_COMPACTION = "IN_MEMORY_COMPACTION"; -060 private static final Bytes IN_MEMORY_COMPACTION_BYTES = new Bytes(Bytes.toBytes(IN_MEMORY_COMPACTION)); -061 -062 @InterfaceAudience.Private -063 public static final String IN_MEMORY = HConstants.IN_MEMORY; -064 private static final Bytes IN_MEMORY_BYTES = new Bytes(Bytes.toBytes(IN_MEMORY)); -065 -066 // These constants are used as FileInfo keys -067 @InterfaceAudience.Private -068 public static final String COMPRESSION = "COMPRESSION"; -069 private static final Bytes COMPRESSION_BYTES = new Bytes(Bytes.toBytes(COMPRESSION)); -070 @InterfaceAudience.Private -071 public static final String COMPRESSION_COMPACT = "COMPRESSION_COMPACT"; -072 private static final Bytes COMPRESSION_COMPACT_BYTES = new Bytes(Bytes.toBytes(COMPRESSION_COMPACT)); -073 @InterfaceAudience.Private -074 public static final String DATA_BLOCK_ENCODING = "DATA_BLOCK_ENCODING"; -075 private static final Bytes DATA_BLOCK_ENCODING_BYTES = new Bytes(Bytes.toBytes(DATA_BLOCK_ENCODING)); -076 /** -077 * Key for the BLOCKCACHE attribute. A more exact name would be -078 * CACHE_DATA_ON_READ because this flag sets whether or not we cache DATA -079 * blocks. We always cache INDEX and BLOOM blocks; caching these blocks cannot -080 * be disabled. -081 */ -082 @InterfaceAudience.Private -083 public static final String BLOCKCACHE = "BLOCKCACHE"; -084 private static final Bytes BLOCKCACHE_BYTES = new Bytes(Bytes.toBytes(BLOCKCACHE)); -085 @InterfaceAudience.Private -086 public static final String CACHE_DATA_ON_WRITE = "CACHE_DATA_ON_WRITE"; -087 private static final Bytes CACHE_DATA_ON_WRITE_BYTES = new Bytes(Bytes.toBytes(CACHE_DATA_ON_WRITE)); -088 @InterfaceAudience.Private -089 public static final String CACHE_INDEX_ON_WRITE = "CACHE_INDEX_ON_WRITE"; -090 private static final Bytes CACHE_INDEX_ON_WRITE_BYTES = new Bytes(Bytes.toBytes(CACHE_INDEX_ON_WRITE)); -091 @InterfaceAudience.Private -092 public static final String CACHE_BLOOMS_ON_WRITE = "CACHE_BLOOMS_ON_WRITE"; -093 private static final Bytes CACHE_BLOOMS_ON_WRITE_BYTES = new Bytes(Bytes.toBytes(CACHE_BLOOMS_ON_WRITE)); -094 @InterfaceAudience.Private -095 public static final String EVICT_BLOCKS_ON_CLOSE = "EVICT_BLOCKS_ON_CLOSE"; -096 private static final Bytes EVICT_BLOCKS_ON_CLOSE_BYTES = new Bytes(Bytes.toBytes(EVICT_BLOCKS_ON_CLOSE)); -097 /** -098
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2d5075d7/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html index afd9ccc..904b921 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html @@ -30,1916 +30,1984 @@ 022import org.apache.hadoop.hbase.shaded.com.google.common.base.Preconditions; 023 024import java.io.IOException; -025import java.util.ArrayList; -026import java.util.Arrays; -027import java.util.HashSet; -028import java.util.Iterator; -029import java.util.List; -030import java.util.Map; -031import java.util.Set; -032import java.util.concurrent.atomic.AtomicBoolean; -033import java.util.concurrent.atomic.AtomicInteger; -034import java.util.concurrent.atomic.AtomicLong; -035import java.util.stream.Collectors; -036import java.util.stream.Stream; -037import java.util.concurrent.ConcurrentHashMap; -038import java.util.concurrent.CopyOnWriteArrayList; -039import java.util.concurrent.DelayQueue; -040import java.util.concurrent.TimeUnit; -041 -042import org.apache.commons.logging.Log; -043import org.apache.commons.logging.LogFactory; -044import org.apache.hadoop.conf.Configuration; -045import org.apache.hadoop.hbase.HConstants; -046import org.apache.hadoop.hbase.ProcedureInfo; -047import org.apache.hadoop.hbase.classification.InterfaceAudience; -048import org.apache.hadoop.hbase.classification.InterfaceStability; -049import org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException; -050import org.apache.hadoop.hbase.procedure2.Procedure.LockState; -051import org.apache.hadoop.hbase.procedure2.store.ProcedureStore; -052import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator; -053import org.apache.hadoop.hbase.procedure2.util.DelayedUtil; -054import org.apache.hadoop.hbase.procedure2.util.DelayedUtil.DelayedWithTimeout; -055import org.apache.hadoop.hbase.procedure2.util.StringUtils; -056import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState; -057import org.apache.hadoop.hbase.security.User; -058import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -059import org.apache.hadoop.hbase.util.NonceKey; -060import org.apache.hadoop.hbase.util.Pair; -061import org.apache.hadoop.hbase.util.Threads; -062 -063/** -064 * Thread Pool that executes the submitted procedures. -065 * The executor has a ProcedureStore associated. -066 * Each operation is logged and on restart the pending procedures are resumed. -067 * -068 * Unless the Procedure code throws an error (e.g. invalid user input) -069 * the procedure will complete (at some point in time), On restart the pending -070 * procedures are resumed and the once failed will be rolledback. -071 * -072 * The user can add procedures to the executor via submitProcedure(proc) -073 * check for the finished state via isFinished(procId) -074 * and get the result via getResult(procId) -075 */ -076@InterfaceAudience.Private -077@InterfaceStability.Evolving -078public class ProcedureExecutorTEnvironment { -079 private static final Log LOG = LogFactory.getLog(ProcedureExecutor.class); -080 -081 public static final String CHECK_OWNER_SET_CONF_KEY = "hbase.procedure.check.owner.set"; -082 private static final boolean DEFAULT_CHECK_OWNER_SET = false; -083 -084 public static final String WORKER_KEEP_ALIVE_TIME_CONF_KEY = -085 "hbase.procedure.worker.keep.alive.time.msec"; -086 private static final long DEFAULT_WORKER_KEEP_ALIVE_TIME = Long.MAX_VALUE; -087 -088 Testing testing = null; -089 public static class Testing { -090protected boolean killIfSuspended = false; -091protected boolean killBeforeStoreUpdate = false; -092protected boolean toggleKillBeforeStoreUpdate = false; -093 -094protected boolean shouldKillBeforeStoreUpdate() { -095 final boolean kill = this.killBeforeStoreUpdate; -096 if (this.toggleKillBeforeStoreUpdate) { -097this.killBeforeStoreUpdate = !kill; -098LOG.warn("Toggle KILL before store update to: " + this.killBeforeStoreUpdate); -099 } -100 return kill; -101} -102 -103protected boolean shouldKillBeforeStoreUpdate(final boolean isSuspended) { -104 return (isSuspended !killIfSuspended) ? false : shouldKillBeforeStoreUpdate(); -105} -106 } -107 -108 public interface ProcedureExecutorListener { -109void procedureLoaded(long procId); -110void procedureAdded(long procId); -111void procedureFinished(long procId); -112 } -113 -114 /**
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0383a9c2/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.html index 15fec0a..d30e78b 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.html @@ -630,684 +630,653 @@ 622} 623 } 624 -625 /** -626 * Sends an MERGE REGIONS RPC to the specified server to merge the specified -627 * regions. -628 * p -629 * A region server could reject the close request because it either does not -630 * have the specified region. -631 * @param server server to merge regions -632 * @param region_a region to merge -633 * @param region_b region to merge -634 * @param forcible true if do a compulsory merge, otherwise we will only merge -635 * two adjacent regions -636 * @throws IOException -637 */ -638 public void sendRegionsMerge(ServerName server, HRegionInfo region_a, -639 HRegionInfo region_b, boolean forcible, final User user) throws IOException { -640if (server == null) -641 throw new NullPointerException("Passed server is null"); -642if (region_a == null || region_b == null) -643 throw new NullPointerException("Passed region is null"); -644AdminService.BlockingInterface admin = getRsAdmin(server); -645if (admin == null) { -646 throw new IOException("Attempting to send MERGE REGIONS RPC to server " -647 + server.toString() + " for region " -648 + region_a.getRegionNameAsString() + "," -649 + region_b.getRegionNameAsString() -650 + " failed because no RPC connection found to this server"); -651} -652HBaseRpcController controller = newRpcController(); -653ProtobufUtil.mergeRegions(controller, admin, region_a, region_b, forcible, user); -654 } -655 -656 @VisibleForTesting -657 public void moveFromOnlineToDeadServers(final ServerName sn) { -658synchronized (onlineServers) { -659 if (!this.onlineServers.containsKey(sn)) { -660LOG.warn("Expiration of " + sn + " but server not online"); -661 } -662 // Remove the server from the known servers lists and update load info BUT -663 // add to deadservers first; do this so it'll show in dead servers list if -664 // not in online servers list. -665 this.deadservers.add(sn); -666 this.onlineServers.remove(sn); -667 onlineServers.notifyAll(); -668} -669this.rsAdmins.remove(sn); -670 } -671 -672 public synchronized void processDeadServer(final ServerName serverName, boolean shouldSplitWal) { -673// When assignment manager is cleaning up the zookeeper nodes and rebuilding the -674// in-memory region states, region servers could be down. Meta table can and -675// should be re-assigned, log splitting can be done too. However, it is better to -676// wait till the cleanup is done before re-assigning user regions. -677// -678// We should not wait in the server shutdown handler thread since it can clog -679// the handler threads and meta table could not be re-assigned in case -680// the corresponding server is down. So we queue them up here instead. -681if (!master.getAssignmentManager().isFailoverCleanupDone()) { -682 requeuedDeadServers.put(serverName, shouldSplitWal); -683 return; -684} -685 -686this.deadservers.add(serverName); -687 master.getAssignmentManager().submitServerCrash(serverName, shouldSplitWal); -688 } -689 -690 /** -691 * Process the servers which died during master's initialization. It will be -692 * called after HMaster#assignMeta and AssignmentManager#joinCluster. -693 * */ -694 synchronized void processQueuedDeadServers() { -695if (!master.isServerCrashProcessingEnabled()) { -696 LOG.info("Master hasn't enabled ServerShutdownHandler"); -697} -698IteratorServerName serverIterator = queuedDeadServers.iterator(); -699while (serverIterator.hasNext()) { -700 ServerName tmpServerName = serverIterator.next(); -701 expireServer(tmpServerName); -702 serverIterator.remove(); -703 requeuedDeadServers.remove(tmpServerName); -704} -705 -706if (!master.getAssignmentManager().isFailoverCleanupDone()) { -707 LOG.info("AssignmentManager hasn't finished failover cleanup; waiting"); -708} -709 -710for (Map.EntryServerName, Boolean entry : requeuedDeadServers.entrySet()) { -711 processDeadServer(entry.getKey(), entry.getValue()); -712} -713requeuedDeadServers.clear(); -714 } -715 -716 /* -717 * Remove the server from the drain list. -718 */ -719 public boolean removeServerFromDrainList(final ServerName sn) { -720// Warn
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f391bcef/devapidocs/org/apache/hadoop/hbase/client/RegionServerCallable.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/RegionServerCallable.html b/devapidocs/org/apache/hadoop/hbase/client/RegionServerCallable.html index 91b6520..f04789a 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/RegionServerCallable.html +++ b/devapidocs/org/apache/hadoop/hbase/client/RegionServerCallable.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":6,"i12":10,"i13":10,"i14":10,"i15":6,"i16":10,"i17":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":6,"i13":10,"i14":10,"i15":10,"i16":6,"i17":10,"i18":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -123,7 +123,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public abstract class RegionServerCallableT,S +public abstract class RegionServerCallableT,S extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements RetryingCallableT Implementations make a RPC call against a RegionService via a protobuf Service. @@ -165,21 +165,25 @@ implements +private int +priority + + private byte[] row - + protected org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController rpcController This is 99% of the time a HBaseRpcController but also used doing Coprocessor Endpoints and in this case, it is a ServerRpcControllable which is not a HBaseRpcController. - + protected S stub - + private TableName tableName @@ -203,6 +207,13 @@ implements +RegionServerCallable(Connectionconnection, +TableNametableName, +byte[]row, + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllerrpcController, +intpriority) + @@ -241,63 +252,67 @@ implements getLocation() +protected int +getPriority() + + byte[] getRow() - + protected org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController getRpcController() - + protected CellScanner getRpcControllerCellScanner() Get the RpcController CellScanner. - + protected S getStub() - + TableName getTableName() - + void prepare(booleanreload) Prepare by setting up any connections to servers, etc., ahead of call invocation. - + protected abstract T rpcCall() Run the RPC call. - + protected void setLocation(HRegionLocationlocation) - + protected void setRpcControllerCellScanner(CellScannercellScanner) - + protected void setStub(Sstub) - + protected abstract void setStubByServiceName(ServerNameserviceName) Set the RCP client stub - + long sleep(longpause, inttries) - + void throwable(http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwablet, booleanretrying) @@ -333,7 +348,7 @@ implements connection -private finalConnection connection +private finalConnection connection @@ -342,7 +357,7 @@ implements tableName -private finalTableName tableName +private finalTableName tableName @@ -351,7 +366,7 @@ implements row -private finalbyte[] row +private finalbyte[] row @@ -360,7 +375,7 @@ implements location -protectedHRegionLocation location +protectedHRegionLocation location Some subclasses want to set their own location. Make it protected. @@ -370,21 +385,30 @@ implements stub -protectedS stub +protectedS stub - + rpcController -protected finalorg.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController rpcController +protected finalorg.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController rpcController This is 99% of the time a HBaseRpcController but also used doing Coprocessor Endpoints and in this case, it is a ServerRpcControllable which is not a HBaseRpcController. Can be null! + + + + + +priority +privateint priority + + @@ -396,10 +420,10 @@ implements - + RegionServerCallable -publicRegionServerCallable(Connectionconnection, +publicRegionServerCallable(Connectionconnection, TableNametableName, byte[]row, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcControllerrpcController) @@ -412,6 +436,19 @@ implements + + + + +RegionServerCallable +publicRegionServerCallable(Connectionconnection, +TableNametableName, +byte[]row, +
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ca5b0275/devapidocs/org/apache/hadoop/hbase/security/visibility/VisibilityReplicationEndpoint.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/security/visibility/VisibilityReplicationEndpoint.html b/devapidocs/org/apache/hadoop/hbase/security/visibility/VisibilityReplicationEndpoint.html index 20d91ce..07607fc 100644 --- a/devapidocs/org/apache/hadoop/hbase/security/visibility/VisibilityReplicationEndpoint.html +++ b/devapidocs/org/apache/hadoop/hbase/security/visibility/VisibilityReplicationEndpoint.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -109,12 +109,12 @@ var activeTableTab = "activeTableTab"; All Implemented Interfaces: -com.google.common.util.concurrent.Service, ReplicationEndpoint, ReplicationPeerConfigListener +ReplicationEndpoint, ReplicationPeerConfigListener, org.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.Service @InterfaceAudience.Private -public class VisibilityReplicationEndpoint +public class VisibilityReplicationEndpoint extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements ReplicationEndpoint @@ -137,11 +137,11 @@ implements ReplicationEndpoint.Context, ReplicationEndpoint.ReplicateContext - + -Nested classes/interfaces inherited from interfacecom.google.common.util.concurrent.Service -com.google.common.util.concurrent.Service.State +Nested classes/interfaces inherited from interfaceorg.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.Service +org.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.Service.Listener, org.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.Service.State @@ -203,65 +203,84 @@ implements Method and Description +void +addListener(org.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.Service.Listenerlistener, + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/Executor.html?is-external=true; title="class or interface in java.util.concurrent">Executorexecutor) + + +void +awaitRunning() + + +void +awaitRunning(longl, +http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true; title="class or interface in java.util.concurrent">TimeUnittimeUnit) + + +void +awaitTerminated() + + +void +awaitTerminated(longl, + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true; title="class or interface in java.util.concurrent">TimeUnittimeUnit) + + boolean canReplicateToSameCluster() Whether or not, the replication endpoint can replicate to it's source cluster with the same UUID - + +http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true; title="class or interface in java.lang">Throwable +failureCause() + + http://docs.oracle.com/javase/8/docs/api/java/util/UUID.html?is-external=true; title="class or interface in java.util">UUID getPeerUUID() Returns a UUID of the provided peer id. - + WALEntryFilter getWALEntryfilter() Returns a WALEntryFilter to use for filtering out WALEntries from the log. - + void init(ReplicationEndpoint.Contextcontext) Initialize the replication endpoint with the given context. - + boolean isRunning() - + void peerConfigUpdated(ReplicationPeerConfigrpc) Callback method for when users update the ReplicationPeerConfig for this peer - + boolean replicate(ReplicationEndpoint.ReplicateContextreplicateContext) Replicate the given set of entries (in the context) to the other cluster. - -com.google.common.util.concurrent.ListenableFuturecom.google.common.util.concurrent.Service.State -start() + +org.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.Service +startAsync() - -com.google.common.util.concurrent.Service.State -startAndWait() - - -com.google.common.util.concurrent.Service.State + +org.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.Service.State state() - -com.google.common.util.concurrent.ListenableFuturecom.google.common.util.concurrent.Service.State -stop() - - -com.google.common.util.concurrent.Service.State -stopAndWait() + +org.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.Service +stopAsync() @@ -291,7 +310,7 @@ implements LOG -private static finalorg.apache.commons.logging.Log LOG +private static
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/9eba7fcf/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmplImpl.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmplImpl.html b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmplImpl.html index 00312b7..962b1ef 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmplImpl.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/tmpl/master/MasterStatusTmplImpl.html @@ -189,318 +189,318 @@ 181AssignmentManager assignmentManager = master.getAssignmentManager(); 182 183// 96, 1 -184jamonWriter.write("!--[if IE]\n!DOCTYPE html\n![endif]--\n?xml version=\"1.0\" encoding=\"UTF-8\" ?\nhtml lang=\"en\"\n head\n meta charset=\"utf-8\"\ntitle"); -185// 103, 12 +184jamonWriter.write("!DOCTYPE html\n?xml version=\"1.0\" encoding=\"UTF-8\" ?\nhtml lang=\"en\"\n head\nmeta charset=\"utf-8\"\n title"); +185// 101, 12 186if (master.isActiveMaster() ) 187{ -188 // 103, 43 +188 // 101, 43 189 jamonWriter.write("Master: "); 190} -191// 103, 51 +191// 101, 51 192else 193{ -194 // 103, 58 +194 // 101, 58 195 jamonWriter.write("Backup Master: "); 196} -197// 103, 79 +197// 101, 79 198jamonWriter.write("\n"); -199// 104, 5 +199// 102, 5 200 org.jamon.escaping.Escaping.NONE.write(org.jamon.emit.StandardEmitter.valueOf(master.getServerName().getHostname()), jamonWriter); -201// 104, 47 +201// 102, 47 202jamonWriter.write("/title\n meta name=\"viewport\" content=\"width=device-width, initial-scale=1.0\"\nmeta name=\"description\" content=\"\"\n link href=\"/static/css/bootstrap.min.css\" rel=\"stylesheet\"\n link href=\"/static/css/bootstrap-theme.min.css\" rel=\"stylesheet\"\n link href=\"/static/css/hbase.css\" rel=\"stylesheet\"\n /head\n\n body\n\ndiv class=\"navbar navbar-fixed-top navbar-default\"\ndiv class=\"container-fluid\"\n div class=\"navbar-header\"\nbutton type=\"button\" class=\"navbar-toggle\" data-toggle=\"collapse\" data-target=\".navbar-collapse\"\nspan class=\"icon-bar\"/span\nspan class=\"icon-bar\"/span\nspan class=\"icon-bar\"/span\n/button t;\na class=\"navbar-brand\" href=\"/master-status\"img src=\"/static/hbase_logo_small.png\" alt=\"HBase Logo\"//a\n/div\n div class=\"collapse navbar-collapse\"\nul class=\"nav navbar-nav\"\nli class=\"active\"a href=\"/\"Home/a/li\nlia href=\"/tablesDetailed.jsp\"Table Details/a/li\n "); -203// 128, 17 +203// 126, 17 204if (master.isActiveMaster() ) 205{ -206 // 128, 48 +206 // 126, 48 207 jamonWriter.write("\n lia href=\"/procedures.jsp\"Procedures amp; Locks/a/li\n"); 208} -209// 130, 23 +209// 128, 23 210jamonWriter.write("\n lia href=\"/processMaster.jsp\"Process Metrics/a/li\nlia href=\"/logs/\"Local Logs/a/li\n lia href=\"/logLevel\"Log Level/a/li\n lia href=\"/dump\"Debug Dump/a/li\n lia href=\"/jmx\"Metrics Dump/a/li\n "); -211// 136, 17 +211// 134, 17 212if (HBaseConfiguration.isShowConfInServlet()) 213{ -214 // 136, 64 +214 // 134, 64 215 jamonWriter.write("\n lia href=\"/conf\"HBase Configuration/a/li\n "); 216} -217// 138, 23 +217// 136, 23 218jamonWriter.write("\n /ul\n/div!--/.nav-collapse --\n /div\n/div\n\ndiv class=\"container-fluid content\"\n\t"); -219// 145, 2 +219// 143, 2 220if (master.isActiveMaster() ) 221{ -222 // 145, 33 +222 // 143, 33 223 jamonWriter.write("\n div class=\"row inner_header\"\ndiv class=\"page-header\"\nh1Master small"); -224 // 148, 35 +224 // 146, 35 225 org.jamon.escaping.Escaping.NONE.write(org.jamon.emit.StandardEmitter.valueOf(master.getServerName().getHostname()), jamonWriter); -226 // 148, 77 +226 // 146, 77 227 jamonWriter.write("/small/h1\n/div\n /div\n\ndiv class=\"row\"\n!-- Various warnings that cluster admins should be aware of --\n"); -228 // 154, 9 +228 // 152, 9 229 if (JvmVersion.isBadJvmVersion() ) 230 { -231// 154, 45 +231// 152, 45 232jamonWriter.write("\n div class=\"alert alert-error\"\n Your current JVM version "); -233// 156, 36
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/17128d27/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html index feb42ea..4bd98f4 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html @@ -185,4189 +185,4266 @@ 177import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest; 178import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest; 179import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse; -180import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest; -181import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest; -182import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse; -183import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest; -184import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos; -185import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos; -186import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse; -187import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos; -188import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils; -189import org.apache.hadoop.hbase.snapshot.HBaseSnapshotException; -190import org.apache.hadoop.hbase.snapshot.RestoreSnapshotException; -191import org.apache.hadoop.hbase.snapshot.SnapshotCreationException; -192import org.apache.hadoop.hbase.snapshot.UnknownSnapshotException; -193import org.apache.hadoop.hbase.util.Addressing; -194import org.apache.hadoop.hbase.util.Bytes; -195import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -196import org.apache.hadoop.hbase.util.ForeignExceptionUtil; -197import org.apache.hadoop.hbase.util.Pair; -198import org.apache.hadoop.hbase.zookeeper.MasterAddressTracker; -199import org.apache.hadoop.hbase.zookeeper.MetaTableLocator; -200import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; -201import org.apache.hadoop.ipc.RemoteException; -202import org.apache.hadoop.util.StringUtils; -203import org.apache.zookeeper.KeeperException; -204 -205import com.google.common.annotations.VisibleForTesting; -206import com.google.protobuf.Descriptors; -207import com.google.protobuf.Message; -208import com.google.protobuf.RpcController; -209import java.util.stream.Collectors; -210 -211/** -212 * HBaseAdmin is no longer a client API. It is marked InterfaceAudience.Private indicating that -213 * this is an HBase-internal class as defined in -214 * https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/InterfaceClassification.html -215 * There are no guarantees for backwards source / binary compatibility and methods or class can -216 * change or go away without deprecation. -217 * Use {@link Connection#getAdmin()} to obtain an instance of {@link Admin} instead of constructing -218 * an HBaseAdmin directly. -219 * -220 * pConnection should be an iunmanaged/i connection obtained via -221 * {@link ConnectionFactory#createConnection(Configuration)} -222 * -223 * @see ConnectionFactory -224 * @see Connection -225 * @see Admin -226 */ -227@InterfaceAudience.Private -228@InterfaceStability.Evolving -229public class HBaseAdmin implements Admin { -230 private static final Log LOG = LogFactory.getLog(HBaseAdmin.class); -231 -232 private static final String ZK_IDENTIFIER_PREFIX = "hbase-admin-on-"; +180import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest; +181import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse; +182import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest; +183import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest; +184import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse; +185import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest; +186import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos; +187import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos; +188import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse; +189import org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos; +190import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils; +191import
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2777c693/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestStore.MyList.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestStore.MyList.html b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestStore.MyList.html index 526f5da..3b51f87 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestStore.MyList.html +++ b/testdevapidocs/org/apache/hadoop/hbase/regionserver/TestStore.MyList.html @@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab"; PrevClass -NextClass +NextClass Frames @@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab"; -private static class TestStore.MyListT +private static class TestStore.MyListT extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListT @@ -143,7 +143,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html delegatee -private http://docs.oracle.com/javase/8/docs/api/java/util/function/Consumer.html?is-external=true; title="class or interface in java.util.function">Consumerhttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer +private TestStore.MyListHook hookAtAdd @@ -161,7 +161,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html Constructor and Description -MyList(http://docs.oracle.com/javase/8/docs/api/java/util/function/Consumer.html?is-external=true; title="class or interface in java.util.function">Consumerhttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">IntegerhookAtAdd) +MyList(TestStore.MyListHookhookAtAdd) @@ -323,7 +323,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html delegatee -private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListT delegatee +private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListT delegatee @@ -332,7 +332,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html hookAtAdd -private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/function/Consumer.html?is-external=true; title="class or interface in java.util.function">Consumerhttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer hookAtAdd +private finalTestStore.MyListHook hookAtAdd @@ -343,13 +343,13 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html Constructor Detail - + MyList -MyList(http://docs.oracle.com/javase/8/docs/api/java/util/function/Consumer.html?is-external=true; title="class or interface in java.util.function">Consumerhttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">IntegerhookAtAdd) +MyList(TestStore.MyListHookhookAtAdd) @@ -366,7 +366,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html size -publicintsize() +publicintsize() Specified by: http://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true#size--; title="class or interface in java.util">sizein interfacehttp://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true; title="class or interface in java.util">CollectionT @@ -381,7 +381,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html isEmpty -publicbooleanisEmpty() +publicbooleanisEmpty() Specified by: http://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true#isEmpty--; title="class or interface in java.util">isEmptyin interfacehttp://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true; title="class or interface in java.util">CollectionT @@ -396,7 +396,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/util/List.html contains -publicbooleancontains(http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Objecto) +publicbooleancontains(http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Objecto) Specified by: http://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true#contains-java.lang.Object-; title="class or interface in java.util">containsin interfacehttp://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true; title="class or interface in
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/90c7dfe4/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html index 75db22d..99a09f9 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html @@ -37,2710 +37,2816 @@ 029import java.util.List; 030import java.util.Map; 031import java.util.Optional; -032import java.util.concurrent.CompletableFuture; -033import java.util.concurrent.TimeUnit; -034import java.util.concurrent.atomic.AtomicReference; -035import java.util.function.BiConsumer; -036import java.util.regex.Pattern; -037import java.util.stream.Collectors; -038 -039import com.google.common.annotations.VisibleForTesting; -040 -041import io.netty.util.Timeout; -042import io.netty.util.TimerTask; -043 -044import java.util.stream.Stream; -045 -046import org.apache.commons.io.IOUtils; -047import org.apache.commons.logging.Log; -048import org.apache.commons.logging.LogFactory; -049import org.apache.hadoop.hbase.ClusterStatus; -050import org.apache.hadoop.hbase.HRegionInfo; -051import org.apache.hadoop.hbase.HRegionLocation; -052import org.apache.hadoop.hbase.MetaTableAccessor; -053import org.apache.hadoop.hbase.MetaTableAccessor.QueryType; -054import org.apache.hadoop.hbase.NotServingRegionException; -055import org.apache.hadoop.hbase.ProcedureInfo; -056import org.apache.hadoop.hbase.RegionLoad; -057import org.apache.hadoop.hbase.RegionLocations; -058import org.apache.hadoop.hbase.ServerName; -059import org.apache.hadoop.hbase.NamespaceDescriptor; -060import org.apache.hadoop.hbase.HConstants; -061import org.apache.hadoop.hbase.TableExistsException; -062import org.apache.hadoop.hbase.TableName; -063import org.apache.hadoop.hbase.AsyncMetaTableAccessor; -064import org.apache.hadoop.hbase.TableNotDisabledException; -065import org.apache.hadoop.hbase.TableNotEnabledException; -066import org.apache.hadoop.hbase.TableNotFoundException; -067import org.apache.hadoop.hbase.UnknownRegionException; -068import org.apache.hadoop.hbase.classification.InterfaceAudience; -069import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder; -070import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder; -071import org.apache.hadoop.hbase.client.Scan.ReadType; -072import org.apache.hadoop.hbase.client.replication.ReplicationSerDeHelper; -073import org.apache.hadoop.hbase.client.replication.TableCFs; -074import org.apache.hadoop.hbase.exceptions.DeserializationException; -075import org.apache.hadoop.hbase.ipc.HBaseRpcController; -076import org.apache.hadoop.hbase.quotas.QuotaFilter; -077import org.apache.hadoop.hbase.quotas.QuotaSettings; -078import org.apache.hadoop.hbase.quotas.QuotaTableUtil; -079import org.apache.hadoop.hbase.replication.ReplicationException; -080import org.apache.hadoop.hbase.replication.ReplicationPeerConfig; -081import org.apache.hadoop.hbase.replication.ReplicationPeerDescription; -082import org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback; -083import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -084import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter; -085import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService; -086import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest; -087import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse; -088import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionRequest; -089import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CompactRegionResponse; -090import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionRequest; -091import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.FlushRegionResponse; -092import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionRequest; -093import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetOnlineRegionResponse; -094import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest; -095import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse; -096import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadRequest; -097import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse; -098import
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0821e51a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html index 71844ce..75db22d 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html @@ -105,2564 +105,2642 @@ 097import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionLoadResponse; 098import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionRequest; 099import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.SplitRegionResponse; -100import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription; -101import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema; -102import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.RegionSpecifierType; -103import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest; -104import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse; -105import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest; -106import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse; -107import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest; -108import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse; -109import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest; -110import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse; -111import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest; -112import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse; -113import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest; -114import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse; -115import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest; -116import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse; -117import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest; -118import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse; -119import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest; -120import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse; -121import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest; -122import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse; -123import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest; -124import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse; -125import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest; -126import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse; -127import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest; -128import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse; -129import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest; -130import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse; -131import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest; -132import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse; -133import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest; -134import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse; -135import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest; -136import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse; -137import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest; -138import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse; -139import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest; -140import
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2d27954a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html index 16c0042..71844ce 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html @@ -126,2499 +126,2543 @@ 118import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse; 119import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest; 120import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse; -121import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest; -122import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse; -123import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest; -124import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse; -125import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest; -126import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse; -127import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest; -128import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse; -129import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest; -130import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse; -131import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest; -132import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse; -133import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest; -134import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse; -135import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest; -136import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse; -137import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest; -138import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse; -139import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest; -140import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse; -141import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest; -142import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse; -143import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest; -144import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse; -145import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest; -146import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse; -147import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest; -148import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse; -149import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest; -150import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse; -151import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest; -152import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse; -153import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest; -154import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse; -155import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest; -156import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse; -157import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest; -158import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse; -159import
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/9fb0764b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html index 369cb85..021cf88 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html @@ -2986,10 +2986,25 @@ service. +(package private) static HTableDescriptor +HBaseAdmin.getHTableDescriptor(TableNametableName, + Connectionconnection, + RpcRetryingCallerFactoryrpcCallerFactory, + RpcControllerFactoryrpcControllerFactory, + intoperationTimeout, + intrpcTimeout) +Deprecated. +since 2.0 version and will be removed in 3.0 version. + use HBaseAdmin.getTableDescriptor(TableName, + Connection, RpcRetryingCallerFactory,RpcControllerFactory,int,int) + + + + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">Optionalhttp://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">Long AsyncHBaseAdmin.getLastMajorCompactionTimestamp(TableNametableName) - + long Admin.getLastMajorCompactionTimestamp(TableNametableName) Get the timestamp of the last major compaction for the passed table @@ -2998,74 +3013,74 @@ service. or 0 if no such HFile could be found. - + long HBaseAdmin.getLastMajorCompactionTimestamp(TableNametableName) - + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">Optionalhttp://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">Long AsyncAdmin.getLastMajorCompactionTimestamp(TableNametableName) Get the timestamp of the last major compaction for the passed table. - + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">Optionalhttp://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">Long RawAsyncHBaseAdmin.getLastMajorCompactionTimestamp(TableNametableName) - + private http://docs.oracle.com/javase/8/docs/api/java/util/function/Consumer.html?is-external=true; title="class or interface in java.util.function">Consumerhttp://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">Long AsyncProcess.getLogger(TableNametableName, longmax) - + private HRegionInfo HBaseAdmin.getMobRegionInfo(TableNametableName) - + int MetaCache.getNumberOfCachedRegionLocations(TableNametableName) Return the number of cached region for a table. - + (package private) int ConnectionImplementation.getNumberOfCachedRegionLocations(TableNametableName) - + default RawAsyncTable AsyncConnection.getRawTable(TableNametableName) Retrieve an RawAsyncTable implementation for accessing a table. - + AsyncTableBuilderRawAsyncTable AsyncConnection.getRawTableBuilder(TableNametableName) Returns an AsyncTableBuilder for creating RawAsyncTable. - + AsyncTableBuilderRawAsyncTable AsyncConnectionImpl.getRawTableBuilder(TableNametableName) - + http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Mapbyte[],RegionLoad Admin.getRegionLoad(ServerNamesn, TableNametableName) Get RegionLoad of all regions hosted on a regionserver for a table. - + http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Mapbyte[],RegionLoad HBaseAdmin.getRegionLoad(ServerNamesn, TableNametableName) - + HRegionLocation ConnectionImplementation.getRegionLocation(TableNametableName, byte[]row, booleanreload) - + HRegionLocation ClusterConnection.getRegionLocation(TableNametableName, byte[]row, @@ -3073,26 +3088,26 @@ service. Find region location hosting passed row - + (package private) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b3b50f22/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/bucket/FileIOEngine.FileWriteAccessor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/bucket/FileIOEngine.FileWriteAccessor.html b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/bucket/FileIOEngine.FileWriteAccessor.html index ba2c5f6..803999d 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/bucket/FileIOEngine.FileWriteAccessor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/bucket/FileIOEngine.FileWriteAccessor.html @@ -60,209 +60,222 @@ 052 private FileReadAccessor readAccessor = new FileReadAccessor(); 053 private FileWriteAccessor writeAccessor = new FileWriteAccessor(); 054 -055 public FileIOEngine(long capacity, String... filePaths) throws IOException { -056this.sizePerFile = capacity / filePaths.length; -057this.capacity = this.sizePerFile * filePaths.length; -058this.filePaths = filePaths; -059this.fileChannels = new FileChannel[filePaths.length]; -060this.rafs = new RandomAccessFile[filePaths.length]; -061for (int i = 0; i filePaths.length; i++) { -062 String filePath = filePaths[i]; -063 try { -064rafs[i] = new RandomAccessFile(filePath, "rw"); -065long totalSpace = new File(filePath).getTotalSpace(); -066if (totalSpace sizePerFile) { -067 // The next setting length will throw exception,logging this message -068 // is just used for the detail reason of exception, -069 String msg = "Only " + StringUtils.byteDesc(totalSpace) -070 + " total space under " + filePath + ", not enough for requested " -071 + StringUtils.byteDesc(sizePerFile); -072 LOG.warn(msg); -073} -074rafs[i].setLength(sizePerFile); -075fileChannels[i] = rafs[i].getChannel(); -076LOG.info("Allocating cache " + StringUtils.byteDesc(sizePerFile) -077+ ", on the path:" + filePath); -078 } catch (IOException fex) { -079LOG.error("Failed allocating cache on " + filePath, fex); -080shutdown(); -081throw fex; -082 } -083} -084 } -085 -086 @Override -087 public String toString() { -088return "ioengine=" + this.getClass().getSimpleName() + ", paths=" -089+ Arrays.asList(filePaths) + ", capacity=" + String.format("%,d", this.capacity); -090 } -091 -092 /** -093 * File IO engine is always able to support persistent storage for the cache -094 * @return true -095 */ -096 @Override -097 public boolean isPersistent() { -098return true; -099 } -100 -101 /** -102 * Transfers data from file to the given byte buffer -103 * @param offset The offset in the file where the first byte to be read -104 * @param length The length of buffer that should be allocated for reading -105 * from the file channel -106 * @return number of bytes read -107 * @throws IOException +055 public FileIOEngine(long capacity, boolean maintainPersistence, String... filePaths) +056 throws IOException { +057this.sizePerFile = capacity / filePaths.length; +058this.capacity = this.sizePerFile * filePaths.length; +059this.filePaths = filePaths; +060this.fileChannels = new FileChannel[filePaths.length]; +061if (!maintainPersistence) { +062 for (String filePath : filePaths) { +063File file = new File(filePath); +064if (file.exists()) { +065 if (LOG.isDebugEnabled()) { +066LOG.debug("File " + filePath + " already exists. Deleting!!"); +067 } +068 file.delete(); +069 // If deletion fails still we can manage with the writes +070} +071 } +072} +073this.rafs = new RandomAccessFile[filePaths.length]; +074for (int i = 0; i filePaths.length; i++) { +075 String filePath = filePaths[i]; +076 try { +077rafs[i] = new RandomAccessFile(filePath, "rw"); +078long totalSpace = new File(filePath).getTotalSpace(); +079if (totalSpace sizePerFile) { +080 // The next setting length will throw exception,logging this message +081 // is just used for the detail reason of exception, +082 String msg = "Only " + StringUtils.byteDesc(totalSpace) +083 + " total space under " + filePath + ", not enough for requested " +084 + StringUtils.byteDesc(sizePerFile); +085 LOG.warn(msg); +086} +087rafs[i].setLength(sizePerFile); +088fileChannels[i] = rafs[i].getChannel(); +089LOG.info("Allocating cache " + StringUtils.byteDesc(sizePerFile) +090+ ", on the path:" + filePath); +091 } catch (IOException fex) { +092LOG.error("Failed allocating cache on " + filePath, fex); +093shutdown();
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ca9f6925/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTable.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTable.html b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTable.html index 1fd28df..ec7f66d 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTable.html +++ b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncTable.html @@ -49,7 +49,7 @@ var activeTableTab = "activeTableTab"; -PrevClass +PrevClass NextClass @@ -393,7 +393,7 @@ extends -PrevClass +PrevClass NextClass http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ca9f6925/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncAdmin.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncAdmin.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncAdmin.html index 3b69c91..d3bf407 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncAdmin.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/AsyncAdmin.html @@ -98,6 +98,27 @@ Uses of AsyncAdmin in org.apache.hadoop.hbase.client +Classes in org.apache.hadoop.hbase.client with type parameters of type AsyncAdmin + +Modifier and Type +Interface and Description + + + +interface +AsyncAdminBuilderT extends AsyncAdmin +For creating AsyncAdmin. + + + +(package private) class +AsyncAdminBuilderBaseT extends AsyncAdmin +Base class for all asynchronous admin builders. + + + + + Classes in org.apache.hadoop.hbase.client that implement AsyncAdmin Modifier and Type @@ -110,6 +131,12 @@ The implementation of AsyncAdmin. + +class +RawAsyncHBaseAdmin +The implementation of AsyncAdmin. + + @@ -121,7 +148,7 @@ protected AsyncAdmin -AsyncHBaseAdmin.ProcedureBiConsumer.admin +RawAsyncHBaseAdmin.ProcedureBiConsumer.admin @@ -133,14 +160,16 @@ -AsyncAdmin +default AsyncAdmin AsyncConnection.getAdmin() -Retrieve an AsyncAdmin implementation to administer an HBase cluster. +Retrieve an AsyncAdmin implementation to administer an HBase cluster. -AsyncAdmin -AsyncConnectionImpl.getAdmin() +default AsyncAdmin +AsyncConnection.getAdmin(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true; title="class or interface in java.util.concurrent">ExecutorServicepool) +Retrieve an AsyncAdmin implementation to administer an HBase cluster. + @@ -151,62 +180,62 @@ -AddColumnFamilyProcedureBiConsumer(AsyncAdminadmin, +AddColumnFamilyProcedureBiConsumer(AsyncAdminadmin, TableNametableName) -CreateNamespaceProcedureBiConsumer(AsyncAdminadmin, +CreateNamespaceProcedureBiConsumer(AsyncAdminadmin, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringnamespaceName) -CreateTableProcedureBiConsumer(AsyncAdminadmin, +CreateTableProcedureBiConsumer(AsyncAdminadmin, TableNametableName) -DeleteColumnFamilyProcedureBiConsumer(AsyncAdminadmin, +DeleteColumnFamilyProcedureBiConsumer(AsyncAdminadmin, TableNametableName) -DeleteNamespaceProcedureBiConsumer(AsyncAdminadmin, +DeleteNamespaceProcedureBiConsumer(AsyncAdminadmin, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringnamespaceName) -DeleteTableProcedureBiConsumer(AsyncAdminadmin, +DeleteTableProcedureBiConsumer(AsyncAdminadmin, TableNametableName) -DisableTableProcedureBiConsumer(AsyncAdminadmin, +DisableTableProcedureBiConsumer(AsyncAdminadmin, TableNametableName) -EnableTableProcedureBiConsumer(AsyncAdminadmin, +EnableTableProcedureBiConsumer(AsyncAdminadmin, TableNametableName) -MergeTableRegionProcedureBiConsumer(AsyncAdminadmin, +MergeTableRegionProcedureBiConsumer(AsyncAdminadmin, TableNametableName) -ModifyColumnFamilyProcedureBiConsumer(AsyncAdminadmin, +ModifyColumnFamilyProcedureBiConsumer(AsyncAdminadmin, TableNametableName) -ModifyNamespaceProcedureBiConsumer(AsyncAdminadmin, +ModifyNamespaceProcedureBiConsumer(AsyncAdminadmin, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringnamespaceName) -NamespaceProcedureBiConsumer(AsyncAdminadmin, +NamespaceProcedureBiConsumer(AsyncAdminadmin, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8e3b63ca/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html index c895448..545d4da 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.CandidateGenerator.html @@ -1294,425 +1294,426 @@ 1286 } 1287 1288 // We normalize locality to be a score between 0 and 1.0 representing how good it -1289 // is compared to how good it could be -1290 locality /= bestLocality; -1291} -1292 -1293@Override -1294protected void regionMoved(int region, int oldServer, int newServer) { -1295 int oldEntity = type == LocalityType.SERVER ? oldServer : cluster.serverIndexToRackIndex[oldServer]; -1296 int newEntity = type == LocalityType.SERVER ? newServer : cluster.serverIndexToRackIndex[newServer]; -1297 if (this.services == null) { -1298return; -1299 } -1300 double localityDelta = getWeightedLocality(region, newEntity) - getWeightedLocality(region, oldEntity); -1301 double normalizedDelta = localityDelta / bestLocality; -1302 locality += normalizedDelta; -1303} -1304 -1305@Override -1306double cost() { -1307 return 1 - locality; -1308} -1309 -1310private int getMostLocalEntityForRegion(int region) { -1311 return cluster.getOrComputeRegionsToMostLocalEntities(type)[region]; -1312} -1313 -1314private double getWeightedLocality(int region, int entity) { -1315 return cluster.getOrComputeWeightedLocality(region, entity, type); -1316} -1317 -1318 } -1319 -1320 static class ServerLocalityCostFunction extends LocalityBasedCostFunction { -1321 -1322private static final String LOCALITY_COST_KEY = "hbase.master.balancer.stochastic.localityCost"; -1323private static final float DEFAULT_LOCALITY_COST = 25; -1324 -1325 ServerLocalityCostFunction(Configuration conf, MasterServices srv) { -1326 super( -1327 conf, -1328 srv, -1329 LocalityType.SERVER, -1330 LOCALITY_COST_KEY, -1331 DEFAULT_LOCALITY_COST -1332 ); -1333} -1334 -1335@Override -1336int regionIndexToEntityIndex(int region) { -1337 return cluster.regionIndexToServerIndex[region]; -1338} -1339 } -1340 -1341 static class RackLocalityCostFunction extends LocalityBasedCostFunction { -1342 -1343private static final String RACK_LOCALITY_COST_KEY = "hbase.master.balancer.stochastic.rackLocalityCost"; -1344private static final float DEFAULT_RACK_LOCALITY_COST = 15; -1345 -1346public RackLocalityCostFunction(Configuration conf, MasterServices services) { -1347 super( -1348 conf, -1349 services, -1350 LocalityType.RACK, -1351 RACK_LOCALITY_COST_KEY, -1352 DEFAULT_RACK_LOCALITY_COST -1353 ); -1354} -1355 -1356@Override -1357int regionIndexToEntityIndex(int region) { -1358 return cluster.getRackForRegion(region); -1359} -1360 } -1361 -1362 /** -1363 * Base class the allows writing costs functions from rolling average of some -1364 * number from RegionLoad. -1365 */ -1366 abstract static class CostFromRegionLoadFunction extends CostFunction { -1367 -1368private ClusterStatus clusterStatus = null; -1369private MapString, DequeBalancerRegionLoad loads = null; -1370private double[] stats = null; -1371 CostFromRegionLoadFunction(Configuration conf) { -1372 super(conf); -1373} -1374 -1375void setClusterStatus(ClusterStatus status) { -1376 this.clusterStatus = status; -1377} -1378 -1379void setLoads(MapString, DequeBalancerRegionLoad l) { -1380 this.loads = l; -1381} -1382 -1383@Override -1384double cost() { -1385 if (clusterStatus == null || loads == null) { -1386return 0; -1387 } -1388 -1389 if (stats == null || stats.length != cluster.numServers) { -1390stats = new double[cluster.numServers]; -1391 } -1392 -1393 for (int i =0; i stats.length; i++) { -1394//Cost this server has from RegionLoad -1395long cost = 0; -1396 -1397// for every region on this server get the rl -1398for(int regionIndex:cluster.regionsPerServer[i]) { -1399 CollectionBalancerRegionLoad regionLoadList = cluster.regionLoads[regionIndex]; -1400 -1401 // Now if we found a region load get the type of cost that was requested. -1402 if (regionLoadList != null) { -1403
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aecb1286/devapidocs/src-html/org/apache/hadoop/hbase/Version.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/Version.html b/devapidocs/src-html/org/apache/hadoop/hbase/Version.html index 114f6d1..a811283 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/Version.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/Version.html @@ -16,11 +16,11 @@ 008@InterfaceAudience.Private 009public class Version { 010 public static final String version = "3.0.0-SNAPSHOT"; -011 public static final String revision = "0e8e176ebd3bd17d969d17ce2b0aa3dafb93fa22"; +011 public static final String revision = "2d781aa15c9535c3845692a275c8babca2c45568"; 012 public static final String user = "jenkins"; -013 public static final String date = "Sun Jun 25 14:39:31 UTC 2017"; +013 public static final String date = "Mon Jun 26 14:39:40 UTC 2017"; 014 public static final String url = "git://asf920.gq1.ygridcore.net/home/jenkins/jenkins-slave/workspace/hbase_generate_website/hbase"; -015 public static final String srcChecksum = "f767e20900e65fc306ff06b652532cf8"; +015 public static final String srcChecksum = "06d75479e61d981c6c5cc84833e9a9c7"; 016}
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a719cd00/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.PrimaryRegionCountSkewCostFunction.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.PrimaryRegionCountSkewCostFunction.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.PrimaryRegionCountSkewCostFunction.html index 6de986f..c895448 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.PrimaryRegionCountSkewCostFunction.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.PrimaryRegionCountSkewCostFunction.html @@ -26,1592 +26,1693 @@ 018package org.apache.hadoop.hbase.master.balancer; 019 020import java.util.ArrayDeque; -021import java.util.Arrays; -022import java.util.Collection; -023import java.util.Deque; -024import java.util.HashMap; -025import java.util.LinkedList; -026import java.util.List; -027import java.util.Map; -028import java.util.Map.Entry; -029import java.util.Random; -030 -031import org.apache.commons.logging.Log; -032import org.apache.commons.logging.LogFactory; -033import org.apache.hadoop.conf.Configuration; -034import org.apache.hadoop.hbase.ClusterStatus; -035import org.apache.hadoop.hbase.HBaseInterfaceAudience; -036import org.apache.hadoop.hbase.HConstants; -037import org.apache.hadoop.hbase.HRegionInfo; -038import org.apache.hadoop.hbase.RegionLoad; -039import org.apache.hadoop.hbase.ServerLoad; -040import org.apache.hadoop.hbase.ServerName; -041import org.apache.hadoop.hbase.TableName; -042import org.apache.hadoop.hbase.classification.InterfaceAudience; -043import org.apache.hadoop.hbase.master.MasterServices; -044import org.apache.hadoop.hbase.master.RegionPlan; -045import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action; -046import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action.Type; -047import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.AssignRegionAction; -048import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.MoveRegionAction; -049import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.SwapRegionsAction; -050import org.apache.hadoop.hbase.util.Bytes; -051import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -052 -053import com.google.common.collect.Lists; -054 -055/** -056 * pThis is a best effort load balancer. Given a Cost function F(C) =gt; x It will -057 * randomly try and mutate the cluster to Cprime. If F(Cprime) lt; F(C) then the -058 * new cluster state becomes the plan. It includes costs functions to compute the cost of:/p -059 * ul -060 * liRegion Load/li -061 * liTable Load/li -062 * liData Locality/li -063 * liMemstore Sizes/li -064 * liStorefile Sizes/li -065 * /ul -066 * -067 * -068 * pEvery cost function returns a number between 0 and 1 inclusive; where 0 is the lowest cost -069 * best solution, and 1 is the highest possible cost and the worst solution. The computed costs are -070 * scaled by their respective multipliers:/p +021import java.util.ArrayList; +022import java.util.Arrays; +023import java.util.Collection; +024import java.util.Collections; +025import java.util.Deque; +026import java.util.HashMap; +027import java.util.LinkedList; +028import java.util.List; +029import java.util.Map; +030import java.util.Map.Entry; +031import java.util.Random; +032 +033import org.apache.commons.logging.Log; +034import org.apache.commons.logging.LogFactory; +035import org.apache.hadoop.conf.Configuration; +036import org.apache.hadoop.hbase.ClusterStatus; +037import org.apache.hadoop.hbase.HBaseInterfaceAudience; +038import org.apache.hadoop.hbase.HConstants; +039import org.apache.hadoop.hbase.HRegionInfo; +040import org.apache.hadoop.hbase.RegionLoad; +041import org.apache.hadoop.hbase.ServerLoad; +042import org.apache.hadoop.hbase.ServerName; +043import org.apache.hadoop.hbase.TableName; +044import org.apache.hadoop.hbase.classification.InterfaceAudience; +045import org.apache.hadoop.hbase.master.MasterServices; +046import org.apache.hadoop.hbase.master.RegionPlan; +047import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action; +048import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action.Type; +049import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.AssignRegionAction; +050import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.LocalityType; +051import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.MoveRegionAction; +052import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.SwapRegionsAction; +053import org.apache.hadoop.hbase.util.Bytes; +054import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; +055
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/476c54ed/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaTableUtil.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaTableUtil.html b/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaTableUtil.html index 509b93c..3c6f9b8 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaTableUtil.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/quotas/QuotaTableUtil.html @@ -53,606 +53,717 @@ 045import org.apache.hadoop.hbase.client.ResultScanner; 046import org.apache.hadoop.hbase.client.Scan; 047import org.apache.hadoop.hbase.client.Table; -048import org.apache.hadoop.hbase.filter.CompareFilter; -049import org.apache.hadoop.hbase.filter.Filter; -050import org.apache.hadoop.hbase.filter.FilterList; -051import org.apache.hadoop.hbase.filter.QualifierFilter; -052import org.apache.hadoop.hbase.filter.RegexStringComparator; -053import org.apache.hadoop.hbase.filter.RowFilter; -054import org.apache.hadoop.hbase.protobuf.ProtobufMagic; -055import org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString; -056import org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException; -057import org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeByteOperations; -058import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -059import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos; -060import org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos; -061import org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesResponse; -062import org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesResponse; -063import org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaSnapshotsResponse; -064import org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaSnapshotsResponse.TableQuotaSnapshot; -065import org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesResponse.RegionSizes; -066import org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas; -067import org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceQuota; -068import org.apache.hadoop.hbase.util.Bytes; -069import org.apache.hadoop.hbase.util.Strings; -070 -071/** -072 * Helper class to interact with the quota table. -073 * pre -074 * ROW-KEY FAM/QUAL DATA -075 * n.lt;namespacegt; q:s lt;global-quotasgt; -076 * t.lt;namespacegt; u:p lt;namespace-quota policygt; -077 * t.lt;tablegt; q:s lt;global-quotasgt; -078 * t.lt;tablegt; u:p lt;table-quota policygt; -079 * u.lt;usergt; q:s lt;global-quotasgt; -080 * u.lt;usergt; q:s.lt;tablegt; lt;table-quotasgt; -081 * u.lt;usergt; q:s.lt;nsgt;: lt;namespace-quotasgt; -082 * /pre -083 */ -084@InterfaceAudience.Private -085@InterfaceStability.Evolving -086public class QuotaTableUtil { -087 private static final Log LOG = LogFactory.getLog(QuotaTableUtil.class); -088 -089 /** System table for quotas */ -090 public static final TableName QUOTA_TABLE_NAME = -091 TableName.valueOf(NamespaceDescriptor.SYSTEM_NAMESPACE_NAME_STR, "quota"); -092 -093 protected static final byte[] QUOTA_FAMILY_INFO = Bytes.toBytes("q"); -094 protected static final byte[] QUOTA_FAMILY_USAGE = Bytes.toBytes("u"); -095 protected static final byte[] QUOTA_QUALIFIER_SETTINGS = Bytes.toBytes("s"); -096 protected static final byte[] QUOTA_QUALIFIER_SETTINGS_PREFIX = Bytes.toBytes("s."); -097 protected static final byte[] QUOTA_QUALIFIER_POLICY = Bytes.toBytes("p"); -098 protected static final String QUOTA_POLICY_COLUMN = -099 Bytes.toString(QUOTA_FAMILY_USAGE) + ":" + Bytes.toString(QUOTA_QUALIFIER_POLICY); -100 protected static final byte[] QUOTA_USER_ROW_KEY_PREFIX = Bytes.toBytes("u."); -101 protected static final byte[] QUOTA_TABLE_ROW_KEY_PREFIX = Bytes.toBytes("t."); -102 protected static final byte[] QUOTA_NAMESPACE_ROW_KEY_PREFIX = Bytes.toBytes("n."); -103 -104 /* = -105 * Quota "settings" helpers -106 */ -107 public static Quotas getTableQuota(final Connection connection, final TableName table) -108 throws IOException { -109return getQuotas(connection, getTableRowKey(table)); -110 } -111 -112 public static Quotas getNamespaceQuota(final Connection connection, final String namespace) -113 throws IOException { -114return getQuotas(connection, getNamespaceRowKey(namespace)); -115 } -116 -117 public static Quotas getUserQuota(final Connection connection, final String user) -118 throws IOException { -119return getQuotas(connection, getUserRowKey(user)); -120 } -121 -122 public static Quotas getUserQuota(final
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/77a552c4/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationState.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationState.html b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationState.html index a969f92..a37ea12 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationState.html +++ b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.ReplicationState.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -private static enum HBaseAdmin.ReplicationState +private static enum HBaseAdmin.ReplicationState extends http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumHBaseAdmin.ReplicationState This enum indicates the current state of the replication for a given table. @@ -214,7 +214,7 @@ the order they are declared. ENABLED -public static finalHBaseAdmin.ReplicationState ENABLED +public static finalHBaseAdmin.ReplicationState ENABLED @@ -223,7 +223,7 @@ the order they are declared. MIXED -public static finalHBaseAdmin.ReplicationState MIXED +public static finalHBaseAdmin.ReplicationState MIXED @@ -232,7 +232,7 @@ the order they are declared. DISABLED -public static finalHBaseAdmin.ReplicationState DISABLED +public static finalHBaseAdmin.ReplicationState DISABLED @@ -249,7 +249,7 @@ the order they are declared. values -public staticHBaseAdmin.ReplicationState[]values() +public staticHBaseAdmin.ReplicationState[]values() Returns an array containing the constants of this enum type, in the order they are declared. This method may be used to iterate over the constants as follows: @@ -269,7 +269,7 @@ for (HBaseAdmin.ReplicationState c : HBaseAdmin.ReplicationState.values()) valueOf -public staticHBaseAdmin.ReplicationStatevalueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) +public staticHBaseAdmin.ReplicationStatevalueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) Returns the enum constant of this type with the specified name. The string must match exactly an identifier used to declare an enum constant in this type. (Extraneous whitespace characters are http://git-wip-us.apache.org/repos/asf/hbase-site/blob/77a552c4/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html index 1417d3a..9bfcda0 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html +++ b/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.RestoreSnapshotFuture.html @@ -127,7 +127,7 @@ var activeTableTab = "activeTableTab"; -private static class HBaseAdmin.RestoreSnapshotFuture +private static class HBaseAdmin.RestoreSnapshotFuture extends HBaseAdmin.TableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void @@ -240,7 +240,7 @@ extends RestoreSnapshotFuture -publicRestoreSnapshotFuture(HBaseAdminadmin, +publicRestoreSnapshotFuture(HBaseAdminadmin, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionsnapshot, TableNametableName, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponseresponse) @@ -252,7 +252,7 @@ extends RestoreSnapshotFuture -publicRestoreSnapshotFuture(HBaseAdminadmin, +publicRestoreSnapshotFuture(HBaseAdminadmin, TableNametableName, http://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">LongprocId) @@ -271,7 +271,7 @@ extends getOperationType -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetOperationType() +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetOperationType() Specified by: getOperationTypein classHBaseAdmin.TableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void http://git-wip-us.apache.org/repos/asf/hbase-site/blob/77a552c4/devapidocs/org/apache/hadoop/hbase/client/HBaseAdmin.TableFuture.TableWaitForStateCallable.html
[18/51] [partial] hbase-site git commit: Published site at 82d554e3783372cc6b05489452c815b57c06f6cd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b44796ef/devapidocs/org/apache/hadoop/hbase/backup/class-use/BackupRestoreConstants.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/backup/class-use/BackupRestoreConstants.html b/devapidocs/org/apache/hadoop/hbase/backup/class-use/BackupRestoreConstants.html index c810478..32b1ef9 100644 --- a/devapidocs/org/apache/hadoop/hbase/backup/class-use/BackupRestoreConstants.html +++ b/devapidocs/org/apache/hadoop/hbase/backup/class-use/BackupRestoreConstants.html @@ -4,7 +4,7 @@ -Uses of Interface org.apache.hadoop.hbase.backup.BackupRestoreConstants (Apache HBase 2.0.0-SNAPSHOT API) +Uses of Interface org.apache.hadoop.hbase.backup.BackupRestoreConstants (Apache HBase 3.0.0-SNAPSHOT API) @@ -12,7 +12,7 @@ -Uses of Class org.apache.hadoop.hbase.backup.BackupRestoreFactory (Apache HBase 2.0.0-SNAPSHOT API) +Uses of Class org.apache.hadoop.hbase.backup.BackupRestoreFactory (Apache HBase 3.0.0-SNAPSHOT API)