[20/51] [partial] hbase-site git commit: Published site at 620d70d6186fb800299bcc62ad7179fccfd1be41.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterWalManager.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterWalManager.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterWalManager.html index 127b088..42abebf 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterWalManager.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterWalManager.html @@ -44,367 +44,369 @@ 036import org.apache.hadoop.hbase.HConstants; 037import org.apache.hadoop.hbase.ServerName; 038import org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL; -039import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -040import org.apache.hadoop.hbase.util.FSUtils; -041import org.apache.hadoop.hbase.wal.AbstractFSWALProvider; -042import org.apache.hadoop.hbase.wal.WALSplitter; -043import org.apache.yetus.audience.InterfaceAudience; -044import org.slf4j.Logger; -045import org.slf4j.LoggerFactory; -046import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; -047 -048/** -049 * This class abstracts a bunch of operations the HMaster needs -050 * when splitting log files e.g. finding log files, dirs etc. -051 */ -052@InterfaceAudience.Private -053public class MasterWalManager { -054 private static final Logger LOG = LoggerFactory.getLogger(MasterWalManager.class); -055 -056 final static PathFilter META_FILTER = new PathFilter() { -057@Override -058public boolean accept(Path p) { -059 return AbstractFSWALProvider.isMetaFile(p); -060} -061 }; -062 -063 @VisibleForTesting -064 public final static PathFilter NON_META_FILTER = new PathFilter() { -065@Override -066public boolean accept(Path p) { -067 return !AbstractFSWALProvider.isMetaFile(p); -068} -069 }; -070 -071 // metrics for master -072 // TODO: Rename it, since those metrics are split-manager related -073 private final MetricsMasterFileSystem metricsMasterFilesystem = new MetricsMasterFileSystem(); -074 -075 // Keep around for convenience. -076 private final MasterServices services; -077 private final Configuration conf; -078 private final FileSystem fs; -079 -080 // The Path to the old logs dir -081 private final Path oldLogDir; -082 private final Path rootDir; -083 -084 // create the split log lock -085 private final Lock splitLogLock = new ReentrantLock(); -086 private final SplitLogManager splitLogManager; -087 -088 // Is the fileystem ok? -089 private volatile boolean fsOk = true; -090 -091 public MasterWalManager(MasterServices services) throws IOException { -092this(services.getConfiguration(), services.getMasterFileSystem().getWALFileSystem(), -093 services.getMasterFileSystem().getWALRootDir(), services); -094 } -095 -096 public MasterWalManager(Configuration conf, FileSystem fs, Path rootDir, MasterServices services) -097 throws IOException { -098this.fs = fs; -099this.conf = conf; -100this.rootDir = rootDir; -101this.services = services; -102this.splitLogManager = new SplitLogManager(services, conf); -103 -104this.oldLogDir = new Path(rootDir, HConstants.HREGION_OLDLOGDIR_NAME); -105 } -106 -107 public void stop() { -108if (splitLogManager != null) { -109 splitLogManager.stop(); -110} -111 } -112 -113 @VisibleForTesting -114 SplitLogManager getSplitLogManager() { -115return this.splitLogManager; -116 } -117 -118 /** -119 * Get the directory where old logs go -120 * @return the dir -121 */ -122 Path getOldLogDir() { -123return this.oldLogDir; -124 } -125 -126 public FileSystem getFileSystem() { -127return this.fs; -128 } -129 -130 /** -131 * Checks to see if the file system is still accessible. -132 * If not, sets closed -133 * @return false if file system is not available -134 */ -135 private boolean checkFileSystem() { -136if (this.fsOk) { -137 try { -138 FSUtils.checkFileSystemAvailable(this.fs); -139 FSUtils.checkDfsSafeMode(this.conf); -140 } catch (IOException e) { -141services.abort("Shutting down HBase cluster: file system not available", e); -142this.fsOk = false; -143 } -144} -145return this.fsOk; -146 } -147 -148 /** -149 * Get Servernames which are currently splitting; paths have a '-splitting' suffix. -150 * @return ServerName -151 * @throws IOException IOException -152 */ -153 public SetServerName getSplittingServersFromWALDir() throws IOException { -154return getServerNamesFromWALDirPath( -155 p - p.getName().endsWith(AbstractFSWALProvider.SPLITTING_EXT)); -156 } -157 -158 /** -159 * Get Servernames that COULD BE 'alive'; excludes those that have a '-splitting' suffix as these -160 * are already being split -- they cannot be 'alive'. -161 * @return ServerName
[20/51] [partial] hbase-site git commit: Published site at 281d6429e55149cc4c05430dcc1d1dc136d8b245.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/901d593a/devapidocs/org/apache/hadoop/hbase/master/MasterServices.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/MasterServices.html b/devapidocs/org/apache/hadoop/hbase/master/MasterServices.html index 03cc2d1..8344ce2 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/MasterServices.html +++ b/devapidocs/org/apache/hadoop/hbase/master/MasterServices.html @@ -18,8 +18,8 @@ catch(err) { } //--> -var methods = {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6,"i20":6,"i21":6,"i22":6,"i23":6,"i24":6,"i25":6,"i26":6,"i27":6,"i28":6,"i29":6,"i30":6,"i31":6,"i32":6,"i33":6,"i34":6,"i35":6,"i36":6,"i37":6,"i38":6,"i39":6,"i40":6,"i41":6,"i42":6,"i43":6,"i44":6,"i45":6,"i46":6,"i47":6,"i48":6,"i49":6,"i50":6,"i51":6,"i52":6,"i53":6,"i54":6,"i55":6,"i56":6,"i57":6,"i58":6}; -var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"]}; +var methods = {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6,"i20":6,"i21":6,"i22":6,"i23":6,"i24":6,"i25":6,"i26":6,"i27":6,"i28":6,"i29":6,"i30":6,"i31":6,"i32":6,"i33":6,"i34":6,"i35":6,"i36":6,"i37":6,"i38":6,"i39":18,"i40":6,"i41":6,"i42":6,"i43":6,"i44":6,"i45":6,"i46":6,"i47":6,"i48":6,"i49":6,"i50":6,"i51":6,"i52":6,"i53":6,"i54":6,"i55":6,"i56":6,"i57":6,"i58":6,"i59":6}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],16:["t5","Default Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; var tableTab = "tableTab"; @@ -129,7 +129,7 @@ extends -All MethodsInstance MethodsAbstract Methods +All MethodsInstance MethodsAbstract MethodsDefault Methods Modifier and Type Method and Description @@ -343,58 +343,62 @@ extends getSnapshotManager() +default SplitWALManager +getSplitWALManager() + + SyncReplicationReplayWALManager getSyncReplicationReplayWALManager() Returns the SyncReplicationReplayWALManager. - + TableDescriptors getTableDescriptors() - + TableStateManager getTableStateManager() - + boolean isActiveMaster() - + boolean isClusterUp() - + boolean isInitialized() - + boolean isInMaintenanceMode() - + boolean isSplitOrMergeEnabled(MasterSwitchTypeswitchType) - + https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListReplicationPeerDescription listReplicationPeers(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringregex) Return a list of replication peers. - + https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableDescriptor listTableDescriptorsByNamespace(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) Get list of table descriptors by namespace - + https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableName listTableNamesByNamespace(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) Get list of table names by namespace - + long mergeRegions(RegionInfo[]regionsToMerge, booleanforcible, @@ -403,7 +407,7 @@ extends Merge regions in a table. - + long modifyColumn(TableNametableName, ColumnFamilyDescriptordescriptor, @@ -412,7 +416,7 @@ extends Modify the column descriptor of an existing column in an existing table - + long modifyTable(TableNametableName, TableDescriptordescriptor, @@ -421,19 +425,19 @@ extends Modify the descriptor of an existing table - + boolean registerService(com.google.protobuf.Serviceinstance) Registers a new protocol buffer Service subclass as a master coprocessor endpoint. - + long removeReplicationPeer(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringpeerId) Removes a peer and stops the replication - + long splitRegion(RegionInforegionInfo, byte[]splitRow, @@ -442,14 +446,14 @@ extends Split a region. - + long transitReplicationPeerSyncReplicationState(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringpeerId, SyncReplicationStateclusterState) Set current cluster state for a synchronous replication peer. - + long truncateTable(TableNametableName,
[20/51] [partial] hbase-site git commit: Published site at 466fa920fee572fe20db3b77ebf539dc304d5f31.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2bf59208/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRpcRetryingCallerFactory.BatchCallerBuilder.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRpcRetryingCallerFactory.BatchCallerBuilder.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRpcRetryingCallerFactory.BatchCallerBuilder.html index 2e150bc..0b315b8 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRpcRetryingCallerFactory.BatchCallerBuilder.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRpcRetryingCallerFactory.BatchCallerBuilder.html @@ -25,22 +25,22 @@ 017 */ 018package org.apache.hadoop.hbase.client; 019 -020import static org.apache.hbase.thirdparty.com.google.common.base.Preconditions.checkArgument; -021import static org.apache.hbase.thirdparty.com.google.common.base.Preconditions.checkNotNull; -022import static org.apache.hadoop.hbase.client.ConnectionUtils.retries2Attempts; +020import static org.apache.hadoop.hbase.client.ConnectionUtils.retries2Attempts; +021import static org.apache.hbase.thirdparty.com.google.common.base.Preconditions.checkArgument; +022import static org.apache.hbase.thirdparty.com.google.common.base.Preconditions.checkNotNull; 023 -024import org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer; -025 -026import java.util.List; -027import java.util.concurrent.CompletableFuture; -028import java.util.concurrent.TimeUnit; -029 -030import org.apache.hadoop.hbase.HRegionLocation; -031import org.apache.hadoop.hbase.ServerName; -032import org.apache.hadoop.hbase.TableName; -033import org.apache.yetus.audience.InterfaceAudience; -034import org.apache.hadoop.hbase.client.metrics.ScanMetrics; -035import org.apache.hadoop.hbase.ipc.HBaseRpcController; +024import java.util.List; +025import java.util.concurrent.CompletableFuture; +026import java.util.concurrent.TimeUnit; +027import org.apache.hadoop.hbase.HRegionLocation; +028import org.apache.hadoop.hbase.ServerName; +029import org.apache.hadoop.hbase.TableName; +030import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +031import org.apache.hadoop.hbase.ipc.HBaseRpcController; +032import org.apache.yetus.audience.InterfaceAudience; +033 +034import org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer; +035 036import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ClientService; 037import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ScanResponse; 038 @@ -83,432 +83,441 @@ 075 076private RegionLocateType locateType = RegionLocateType.CURRENT; 077 -078public SingleRequestCallerBuilderT table(TableName tableName) { -079 this.tableName = tableName; -080 return this; -081} -082 -083public SingleRequestCallerBuilderT row(byte[] row) { -084 this.row = row; -085 return this; -086} -087 -088public SingleRequestCallerBuilderT action( -089 AsyncSingleRequestRpcRetryingCaller.CallableT callable) { -090 this.callable = callable; -091 return this; -092} -093 -094public SingleRequestCallerBuilderT operationTimeout(long operationTimeout, TimeUnit unit) { -095 this.operationTimeoutNs = unit.toNanos(operationTimeout); -096 return this; -097} -098 -099public SingleRequestCallerBuilderT rpcTimeout(long rpcTimeout, TimeUnit unit) { -100 this.rpcTimeoutNs = unit.toNanos(rpcTimeout); -101 return this; -102} -103 -104public SingleRequestCallerBuilderT locateType(RegionLocateType locateType) { -105 this.locateType = locateType; -106 return this; -107} -108 -109public SingleRequestCallerBuilderT pause(long pause, TimeUnit unit) { -110 this.pauseNs = unit.toNanos(pause); -111 return this; -112} -113 -114public SingleRequestCallerBuilderT maxAttempts(int maxAttempts) { -115 this.maxAttempts = maxAttempts; -116 return this; -117} -118 -119public SingleRequestCallerBuilderT startLogErrorsCnt(int startLogErrorsCnt) { -120 this.startLogErrorsCnt = startLogErrorsCnt; -121 return this; -122} -123 -124public AsyncSingleRequestRpcRetryingCallerT build() { -125 return new AsyncSingleRequestRpcRetryingCaller(retryTimer, conn, -126 checkNotNull(tableName, "tableName is null"), checkNotNull(row, "row is null"), -127 checkNotNull(locateType, "locateType is null"), checkNotNull(callable, "action is null"), -128 pauseNs, maxAttempts, operationTimeoutNs, rpcTimeoutNs, startLogErrorsCnt); +078private int replicaId = RegionReplicaUtil.DEFAULT_REPLICA_ID; +079 +080public SingleRequestCallerBuilderT table(TableName tableName) { +081 this.tableName = tableName; +082 return this; +083} +084 +085public SingleRequestCallerBuilderT row(byte[] row) { +086 this.row = row; +087
[20/51] [partial] hbase-site git commit: Published site at e4b6b4afb933a961f543537875f87a2dc62d3757.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftHBaseServiceHandler.ResultScannerWrapper.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftHBaseServiceHandler.ResultScannerWrapper.html b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftHBaseServiceHandler.ResultScannerWrapper.html new file mode 100644 index 000..5b5b199 --- /dev/null +++ b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftHBaseServiceHandler.ResultScannerWrapper.html @@ -0,0 +1,1419 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + +Source code + + + + +001/** +002 * +003 * Licensed to the Apache Software Foundation (ASF) under one +004 * or more contributor license agreements. See the NOTICE file +005 * distributed with this work for additional information +006 * regarding copyright ownership. The ASF licenses this file +007 * to you under the Apache License, Version 2.0 (the +008 * "License"); you may not use this file except in compliance +009 * with the License. You may obtain a copy of the License at +010 * +011 * http://www.apache.org/licenses/LICENSE-2.0 +012 * +013 * Unless required by applicable law or agreed to in writing, software +014 * distributed under the License is distributed on an "AS IS" BASIS, +015 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +016 * See the License for the specific language governing permissions and +017 * limitations under the License. +018 */ +019 +020package org.apache.hadoop.hbase.thrift; +021 +022import static org.apache.hadoop.hbase.thrift.Constants.COALESCE_INC_KEY; +023import static org.apache.hadoop.hbase.util.Bytes.getBytes; +024 +025import java.io.IOException; +026import java.nio.ByteBuffer; +027import java.util.ArrayList; +028import java.util.Collections; +029import java.util.HashMap; +030import java.util.List; +031import java.util.Map; +032import java.util.TreeMap; +033 +034import org.apache.hadoop.conf.Configuration; +035import org.apache.hadoop.hbase.Cell; +036import org.apache.hadoop.hbase.CellBuilder; +037import org.apache.hadoop.hbase.CellBuilderFactory; +038import org.apache.hadoop.hbase.CellBuilderType; +039import org.apache.hadoop.hbase.CellUtil; +040import org.apache.hadoop.hbase.HColumnDescriptor; +041import org.apache.hadoop.hbase.HConstants; +042import org.apache.hadoop.hbase.HRegionLocation; +043import org.apache.hadoop.hbase.HTableDescriptor; +044import org.apache.hadoop.hbase.KeyValue; +045import org.apache.hadoop.hbase.MetaTableAccessor; +046import org.apache.hadoop.hbase.ServerName; +047import org.apache.hadoop.hbase.TableName; +048import org.apache.hadoop.hbase.TableNotFoundException; +049import org.apache.hadoop.hbase.client.Append; +050import org.apache.hadoop.hbase.client.Delete; +051import org.apache.hadoop.hbase.client.Durability; +052import org.apache.hadoop.hbase.client.Get; +053import org.apache.hadoop.hbase.client.Increment; +054import org.apache.hadoop.hbase.client.OperationWithAttributes; +055import org.apache.hadoop.hbase.client.Put; +056import org.apache.hadoop.hbase.client.RegionInfo; +057import org.apache.hadoop.hbase.client.RegionLocator; +058import org.apache.hadoop.hbase.client.Result; +059import org.apache.hadoop.hbase.client.ResultScanner; +060import org.apache.hadoop.hbase.client.Scan; +061import org.apache.hadoop.hbase.client.Table; +062import org.apache.hadoop.hbase.filter.Filter; +063import org.apache.hadoop.hbase.filter.ParseFilter; +064import org.apache.hadoop.hbase.filter.PrefixFilter; +065import org.apache.hadoop.hbase.filter.WhileMatchFilter; +066import org.apache.hadoop.hbase.security.UserProvider; +067import org.apache.hadoop.hbase.thrift.generated.AlreadyExists; +068import org.apache.hadoop.hbase.thrift.generated.BatchMutation; +069import org.apache.hadoop.hbase.thrift.generated.ColumnDescriptor; +070import org.apache.hadoop.hbase.thrift.generated.Hbase; +071import org.apache.hadoop.hbase.thrift.generated.IOError; +072import org.apache.hadoop.hbase.thrift.generated.IllegalArgument; +073import org.apache.hadoop.hbase.thrift.generated.Mutation; +074import org.apache.hadoop.hbase.thrift.generated.TAppend; +075import org.apache.hadoop.hbase.thrift.generated.TCell; +076import org.apache.hadoop.hbase.thrift.generated.TIncrement; +077import org.apache.hadoop.hbase.thrift.generated.TRegionInfo; +078import org.apache.hadoop.hbase.thrift.generated.TRowResult; +079import org.apache.hadoop.hbase.thrift.generated.TScan; +080import org.apache.hadoop.hbase.util.Bytes; +081import org.apache.thrift.TException; +082import org.apache.yetus.audience.InterfaceAudience; +083import org.slf4j.Logger; +084import org.slf4j.LoggerFactory; +085 +086import org.apache.hbase.thirdparty.com.google.common.base.Throwables; +087 +088/** +089 * The HBaseServiceHandler is a glue object that
[20/51] [partial] hbase-site git commit: Published site at 3ab895979b643a2980bcdb7fee2078f14b614210.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/exceptions/package-use.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/exceptions/package-use.html b/devapidocs/org/apache/hadoop/hbase/exceptions/package-use.html index d51f5e6..555dae3 100644 --- a/devapidocs/org/apache/hadoop/hbase/exceptions/package-use.html +++ b/devapidocs/org/apache/hadoop/hbase/exceptions/package-use.html @@ -566,6 +566,6 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/executor/EventHandler.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/executor/EventHandler.html b/devapidocs/org/apache/hadoop/hbase/executor/EventHandler.html index bead54b..cbdc6ff 100644 --- a/devapidocs/org/apache/hadoop/hbase/executor/EventHandler.html +++ b/devapidocs/org/apache/hadoop/hbase/executor/EventHandler.html @@ -619,6 +619,6 @@ implements https://docs.oracle.com/javase/8/docs/api/java/lang/Runnable -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/executor/EventType.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/executor/EventType.html b/devapidocs/org/apache/hadoop/hbase/executor/EventType.html index 9a7165e..c3ff4e0 100644 --- a/devapidocs/org/apache/hadoop/hbase/executor/EventType.html +++ b/devapidocs/org/apache/hadoop/hbase/executor/EventType.html @@ -1183,6 +1183,6 @@ not permitted.) -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/executor/ExecutorService.Executor.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/executor/ExecutorService.Executor.html b/devapidocs/org/apache/hadoop/hbase/executor/ExecutorService.Executor.html index 0daf97d..8290699 100644 --- a/devapidocs/org/apache/hadoop/hbase/executor/ExecutorService.Executor.html +++ b/devapidocs/org/apache/hadoop/hbase/executor/ExecutorService.Executor.html @@ -431,6 +431,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/executor/ExecutorService.ExecutorStatus.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/executor/ExecutorService.ExecutorStatus.html b/devapidocs/org/apache/hadoop/hbase/executor/ExecutorService.ExecutorStatus.html index cbfa071..c603d8c 100644 --- a/devapidocs/org/apache/hadoop/hbase/executor/ExecutorService.ExecutorStatus.html +++ b/devapidocs/org/apache/hadoop/hbase/executor/ExecutorService.ExecutorStatus.html @@ -357,6 +357,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html b/devapidocs/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html index e4289ac..64d5fcb 100644 --- a/devapidocs/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html +++ b/devapidocs/org/apache/hadoop/hbase/executor/ExecutorService.RunningEventStatus.html @@ -290,6 +290,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache
[20/51] [partial] hbase-site git commit: Published site at 7820ba1dbdba58b1002cdfde08eb21aa7a0bb6da.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/tool/WriteSinkCoprocessor.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/tool/WriteSinkCoprocessor.html b/devapidocs/org/apache/hadoop/hbase/tool/WriteSinkCoprocessor.html index 938c09f..bd49ba8 100644 --- a/devapidocs/org/apache/hadoop/hbase/tool/WriteSinkCoprocessor.html +++ b/devapidocs/org/apache/hadoop/hbase/tool/WriteSinkCoprocessor.html @@ -278,7 +278,7 @@ implements RegionObserver -postAppend, postBatchMutate, postBatchMutateIndispensably, postBulkLoadHFile, postCheckAndDelete, postCheckAndPut, postClose, postCloseRegionOperation, postCommitStoreFile, postCompact, postCompactSelection, postDelete, postExists, postFlush, postFlush, postGetOp, postIncrement, postInstantiateDeleteTracker, postMemStoreCompacti on, postMutationBeforeWAL, postOpen, postPut, po stReplayWALs, postScannerClose, postScannerFilterRow, postScannerNext, postScannerOpen, postStartRegionOperation, postStoreFileReaderOpen, postWALRestore, preAppend, preAppendAfterRowLock, preBulkLoadHFile, preCheckAndDelete, preCheckAndDeleteAfterRowLock, preCheckAndPut, preCheckAndPutAfterRowLock, preClose, preCommitStoreFile, preCompact, preCompactScannerOpen, preCompactSelection, preDelete, preExists, preFlush, preFlush, preFlushScannerOpen, preGetOp, preIncrement, preIncrementAfterRowLock, preMemStoreCompaction, preMemStoreCompactionCompact, preMemStoreCompactionCompactScannerOpen, prePrepareTimeStampForDeleteVersion, prePut, preReplayWALs, preScannerClose, preScannerNext, preScannerOpen, preStoreFileReaderOpen, preStoreScannerOpen, preWALRestore +postAppend, postAppendBeforeWAL, postBatchMutate, postBatchMutateIndispensably, postBulkLoadHFile, postCheckAndDelete, postCheckAndPut, postClose, postCloseRegionOperation, postCommitStoreFile, postCompact, postCompactSelection, postDelete, postExists, postFlush, postFlush, postGetOp, postIncrement, postIncrementBeforeWAL, postInstantiateDeleteTracker, postMemStoreCompaction, postMutationBeforeWAL, postOpen, postPut, postReplayWALs, postScannerClose, postScannerFilterRow, postScannerNext, postScannerOpen, postStartRegionOperation, postStoreFileReaderOpen, postWALRestore, preAppend, preAppendAfterRowLock, preBulkLoadHFile, preCheckAndDelete, preCheckAndDeleteAfterRowLock, preCheckAndPut, preCheckAndPutAfterRowLock, preClose, preCommitStoreFile, preCompact, preCompactScannerOpen, preCompactSelection, preDelete, preExists, preFlush, preFlush, preFlushScannerOpen, preGetOp, preIncrement, preIncrementA fterRowLock, preMemStoreCompaction, preMemStoreCompactionCompact, preMemStoreCompactionCompactScannerOpen, prePrepareTimeStampForDeleteVersion, prePut, preReplayWALs, preScannerClose, preScannerNext, preScannerOpen, preStoreFileReaderOpen, preStoreScannerOpen, preWALRestore http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html b/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html index da375bf..9db284e 100644 --- a/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html +++ b/devapidocs/org/apache/hadoop/hbase/util/class-use/Pair.html @@ -675,6 +675,33 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. Uses of Pair in org.apache.hadoop.hbase.coprocessor +Methods in org.apache.hadoop.hbase.coprocessor that return types with arguments of type Pair + +Modifier and Type +Method and Description + + + +default https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListPairCell,Cell +RegionObserver.postAppendBeforeWAL(ObserverContextRegionCoprocessorEnvironmentctx, + Mutationmutation, + https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListPairCell,CellcellPairs) +Called after a list of new cells has been created during an append operation, but before + they are committed to the WAL or memstore. + + + +default https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListPairCell,Cell +RegionObserver.postIncrementBeforeWAL(ObserverContextRegionCoprocessorEnvironmentctx, + Mutationmutation, + https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListPairCell,CellcellPairs) +Called after a list of new cells has been created during an increment operation, but before + they are committed to the WAL or memstore. + + + + + Method
[20/51] [partial] hbase-site git commit: Published site at c448604ceb987d113913f0583452b2abce04db0d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html b/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html index 034077c..c20ff47 100644 --- a/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/rest/model/package-tree.html @@ -110,8 +110,8 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.rest.model.ScannerModel.FilterModel.ByteArrayComparableModel.ComparatorType org.apache.hadoop.hbase.rest.model.ScannerModel.FilterModel.FilterType +org.apache.hadoop.hbase.rest.model.ScannerModel.FilterModel.ByteArrayComparableModel.ComparatorType http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html b/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html index aa0564c..56f5ff2 100644 --- a/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/security/access/package-tree.html @@ -141,10 +141,10 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.security.access.Permission.Scope -org.apache.hadoop.hbase.security.access.AccessController.OpType -org.apache.hadoop.hbase.security.access.Permission.Action org.apache.hadoop.hbase.security.access.AccessControlFilter.Strategy +org.apache.hadoop.hbase.security.access.Permission.Action +org.apache.hadoop.hbase.security.access.AccessController.OpType +org.apache.hadoop.hbase.security.access.Permission.Scope http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html index 2cef8bd..4ade4c1 100644 --- a/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/thrift/package-tree.html @@ -199,8 +199,8 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.thrift.ThriftServerRunner.ImplType org.apache.hadoop.hbase.thrift.ThriftMetrics.ThriftServerType +org.apache.hadoop.hbase.thrift.ThriftServerRunner.ImplType org.apache.hadoop.hbase.thrift.MetricsThriftServerSourceFactoryImpl.FactoryStorage http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/BlockCacheTmpl.ImplData.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/BlockCacheTmpl.ImplData.html b/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/BlockCacheTmpl.ImplData.html index c0d1f12..20b038d 100644 --- a/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/BlockCacheTmpl.ImplData.html +++ b/devapidocs/org/apache/hadoop/hbase/tmpl/regionserver/BlockCacheTmpl.ImplData.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab"; -public static class BlockCacheTmpl.ImplData
[20/51] [partial] hbase-site git commit: Published site at 8bf966c8e936dec4d83bcbe85c5aab543f14a0df.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/27555316/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.html b/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.html index 809f66f..9b60dd0 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.html @@ -765,146 +765,145 @@ 757found.set(true); 758try { 759 boolean rootMetaFound = -760 masterServices.getMetaTableLocator().verifyMetaRegionLocation( -761 conn, masterServices.getZooKeeper(), 1); -762 if (rootMetaFound) { -763MetaTableAccessor.Visitor visitor = new DefaultVisitorBase() { -764 @Override -765 public boolean visitInternal(Result row) throws IOException { -766RegionInfo info = MetaTableAccessor.getRegionInfo(row); -767if (info != null) { -768 Cell serverCell = -769 row.getColumnLatestCell(HConstants.CATALOG_FAMILY, -770 HConstants.SERVER_QUALIFIER); -771 if (RSGROUP_TABLE_NAME.equals(info.getTable()) serverCell != null) { -772ServerName sn = -773 ServerName.parseVersionedServerName(CellUtil.cloneValue(serverCell)); -774if (sn == null) { -775 found.set(false); -776} else if (tsm.isTableState(RSGROUP_TABLE_NAME, TableState.State.ENABLED)) { -777 try { -778 ClientProtos.ClientService.BlockingInterface rs = conn.getClient(sn); -779 ClientProtos.GetRequest request = -780 RequestConverter.buildGetRequest(info.getRegionName(), -781new Get(ROW_KEY)); -782rs.get(null, request); -783 assignedRegions.add(info); -784 } catch(Exception ex) { -785LOG.debug("Caught exception while verifying group region", ex); -786 } -787} -788 foundRegions.add(info); -789 } -790} -791return true; -792 } -793}; -794 MetaTableAccessor.fullScanRegions(conn, visitor); -795// if no regions in meta then we have to create the table -796if (foundRegions.size() 1 rootMetaFound !createSent) { -797 createRSGroupTable(); -798 createSent = true; -799} -800LOG.info("RSGroup table=" + RSGROUP_TABLE_NAME + " isOnline=" + found.get() -801+ ", regionCount=" + foundRegions.size() + ", assignCount=" -802+ assignedRegions.size() + ", rootMetaFound=" + rootMetaFound); -803found.set(found.get() assignedRegions.size() == foundRegions.size() -804 foundRegions.size() 0); -805 } else { -806LOG.info("Waiting for catalog tables to come online"); -807found.set(false); -808 } -809 if (found.get()) { -810LOG.debug("With group table online, refreshing cached information."); -811 RSGroupInfoManagerImpl.this.refresh(true); -812online = true; -813//flush any inconsistencies between ZK and HTable -814 RSGroupInfoManagerImpl.this.flushConfig(); -815 } -816} catch (RuntimeException e) { -817 throw e; -818} catch(Exception e) { -819 found.set(false); -820 LOG.warn("Failed to perform check", e); -821} -822try { -823 Thread.sleep(100); -824} catch (InterruptedException e) { -825 LOG.info("Sleep interrupted", e); -826} -827 } -828 return found.get(); -829} -830 -831private void createRSGroupTable() throws IOException { -832 Long procId = masterServices.createSystemTable(RSGROUP_TABLE_DESC); -833 // wait for region to be online -834 int tries = 600; -835 while (!(masterServices.getMasterProcedureExecutor().isFinished(procId)) -836 masterServices.getMasterProcedureExecutor().isRunning() -837 tries 0) { -838try { -839 Thread.sleep(100); -840} catch (InterruptedException e) { -841 throw new IOException("Wait interrupted ", e); -842} -843tries--; -844 } -845 if(tries = 0) { -846throw new IOException("Failed to create group table in a given time."); -847 } else { -848Procedure? result =
[20/51] [partial] hbase-site git commit: Published site at 1acbd36c903b048141866b143507bfce124a5c5f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedureSyncWait.Predicate.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedureSyncWait.Predicate.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedureSyncWait.Predicate.html index 897da76..20face7 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedureSyncWait.Predicate.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedureSyncWait.Predicate.html @@ -112,6 +112,13 @@ static TT +ProcedureSyncWait.waitFor(MasterProcedureEnvenv, + longwaitTime, + https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringpurpose, + ProcedureSyncWait.PredicateTpredicate) + + +static TT ProcedureSyncWait.waitFor(MasterProcedureEnvenv, https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringpurpose, ProcedureSyncWait.PredicateTpredicate) http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html index 2991c16..1dd5d5a 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html @@ -216,8 +216,8 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) org.apache.hadoop.hbase.master.procedure.TableProcedureInterface.TableOperationType -org.apache.hadoop.hbase.master.procedure.ServerProcedureInterface.ServerOperationType org.apache.hadoop.hbase.master.procedure.PeerProcedureInterface.PeerOperationType +org.apache.hadoop.hbase.master.procedure.ServerProcedureInterface.ServerOperationType org.apache.hadoop.hbase.master.procedure.MetaProcedureInterface.MetaOperationType http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.RefreshCacheTask.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.RefreshCacheTask.html b/devapidocs/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.RefreshCacheTask.html index 8923097..499e7bb 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.RefreshCacheTask.html +++ b/devapidocs/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.RefreshCacheTask.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab"; -public class SnapshotFileCache.RefreshCacheTask +public class SnapshotFileCache.RefreshCacheTask extends https://docs.oracle.com/javase/8/docs/api/java/util/TimerTask.html?is-external=true; title="class or interface in java.util">TimerTask Simple helper task that just periodically attempts to refresh the cache @@ -199,7 +199,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/util/TimerTask.h RefreshCacheTask -publicRefreshCacheTask() +publicRefreshCacheTask() @@ -216,7 +216,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/util/TimerTask.h run -publicvoidrun() +publicvoidrun() Specified by: https://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true#run--; title="class or interface in java.lang">runin interfacehttps://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true; title="class or interface in java.lang">Runnable http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.SnapshotDirectoryInfo.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.SnapshotDirectoryInfo.html b/devapidocs/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.SnapshotDirectoryInfo.html index af18761..171d0c9 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.SnapshotDirectoryInfo.html +++ b/devapidocs/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.SnapshotDirectoryInfo.html @@
[20/51] [partial] hbase-site git commit: Published site at 130057f13774f6b213cdb06952c805a29d59396e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/68eae623/devapidocs/src-html/org/apache/hadoop/hbase/security/access/AuthManager.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/security/access/AuthManager.html b/devapidocs/src-html/org/apache/hadoop/hbase/security/access/AuthManager.html new file mode 100644 index 000..4d5cbc9 --- /dev/null +++ b/devapidocs/src-html/org/apache/hadoop/hbase/security/access/AuthManager.html @@ -0,0 +1,680 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + +Source code + + + + +001/** +002 * Licensed to the Apache Software Foundation (ASF) under one +003 * or more contributor license agreements. See the NOTICE file +004 * distributed with this work for additional information +005 * regarding copyright ownership. The ASF licenses this file +006 * to you under the Apache License, Version 2.0 (the +007 * "License"); you may not use this file except in compliance +008 * with the License. You may obtain a copy of the License at +009 * +010 * http://www.apache.org/licenses/LICENSE-2.0 +011 * +012 * Unless required by applicable law or agreed to in writing, software +013 * distributed under the License is distributed on an "AS IS" BASIS, +014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +015 * See the License for the specific language governing permissions and +016 * limitations under the License. +017 */ +018 +019package org.apache.hadoop.hbase.security.access; +020 +021import java.io.Closeable; +022import java.io.IOException; +023import java.util.HashMap; +024import java.util.HashSet; +025import java.util.List; +026import java.util.Map; +027import java.util.Set; +028import java.util.concurrent.ConcurrentHashMap; +029import java.util.concurrent.atomic.AtomicLong; +030 +031import org.apache.hadoop.conf.Configuration; +032import org.apache.hadoop.hbase.AuthUtil; +033import org.apache.hadoop.hbase.Cell; +034import org.apache.hadoop.hbase.TableName; +035import org.apache.hadoop.hbase.exceptions.DeserializationException; +036import org.apache.hadoop.hbase.log.HBaseMarkers; +037import org.apache.hadoop.hbase.security.Superusers; +038import org.apache.hadoop.hbase.security.User; +039import org.apache.hadoop.hbase.security.UserProvider; +040import org.apache.hadoop.hbase.util.Bytes; +041import org.apache.hadoop.hbase.zookeeper.ZKWatcher; +042import org.apache.yetus.audience.InterfaceAudience; +043import org.apache.zookeeper.KeeperException; +044import org.slf4j.Logger; +045import org.slf4j.LoggerFactory; +046 +047import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; +048import org.apache.hbase.thirdparty.com.google.common.collect.ListMultimap; +049import org.apache.hbase.thirdparty.com.google.common.collect.Lists; +050 +051/** +052 * Performs authorization checks for a given user's assigned permissions. +053 * p +054 * There're following scopes: bGlobal/b, bNamespace/b, bTable/b, bFamily/b, +055 * bQualifier/b, bCell/b. +056 * Generally speaking, higher scopes can overrides lower scopes, +057 * except for Cell permission can be granted even a user has not permission on specified table, +058 * which means the user can get/scan only those granted cells parts. +059 * /p +060 * e.g, if user A has global permission R(ead), he can +061 * read table T without checking table scope permission, so authorization checks alway starts from +062 * Global scope. +063 * p +064 * For each scope, not only user but also groups he belongs to will be checked. +065 * /p +066 */ +067@InterfaceAudience.Private +068public final class AuthManager implements Closeable { +069 +070 /** +071 * Cache of permissions, it is thread safe. +072 * @param T T extends Permission +073 */ +074 private static class PermissionCacheT extends Permission { +075private final Object mutex = new Object(); +076private MapString, SetT cache = new HashMap(); +077 +078void put(String name, T perm) { +079 synchronized (mutex) { +080SetT perms = cache.getOrDefault(name, new HashSet()); +081perms.add(perm); +082cache.put(name, perms); +083 } +084} +085 +086SetT get(String name) { +087 synchronized (mutex) { +088return cache.get(name); +089 } +090} +091 +092void clear() { +093 synchronized (mutex) { +094for (Map.EntryString, SetT entry : cache.entrySet()) { +095 entry.getValue().clear(); +096} +097cache.clear(); +098 } +099} +100 } +101 PermissionCacheNamespacePermission NS_NO_PERMISSION = new PermissionCache(); +102 PermissionCacheTablePermission TBL_NO_PERMISSION = new PermissionCache(); +103 +104 /** +105 * Cache for global permission. +106 * Since every user/group can only have one global permission, no need to user PermissionCache. +107 */ +108 private volatile MapString,
[20/51] [partial] hbase-site git commit: Published site at d5e4faacc354c1bc4d93efa71ca97ee3a056123e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b5e107c3/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.PeriodicMemStoreFlusher.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.PeriodicMemStoreFlusher.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.PeriodicMemStoreFlusher.html index b2a9771..bf81ebb 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.PeriodicMemStoreFlusher.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.PeriodicMemStoreFlusher.html @@ -46,3768 +46,3806 @@ 038import java.util.Objects; 039import java.util.Set; 040import java.util.SortedMap; -041import java.util.TreeMap; -042import java.util.TreeSet; -043import java.util.concurrent.ConcurrentHashMap; -044import java.util.concurrent.ConcurrentMap; -045import java.util.concurrent.ConcurrentSkipListMap; -046import java.util.concurrent.atomic.AtomicBoolean; -047import java.util.concurrent.locks.ReentrantReadWriteLock; -048import java.util.function.Function; -049import javax.management.MalformedObjectNameException; -050import javax.servlet.http.HttpServlet; -051import org.apache.commons.lang3.RandomUtils; -052import org.apache.commons.lang3.StringUtils; -053import org.apache.commons.lang3.SystemUtils; -054import org.apache.hadoop.conf.Configuration; -055import org.apache.hadoop.fs.FileSystem; -056import org.apache.hadoop.fs.Path; -057import org.apache.hadoop.hbase.Abortable; -058import org.apache.hadoop.hbase.CacheEvictionStats; -059import org.apache.hadoop.hbase.ChoreService; -060import org.apache.hadoop.hbase.ClockOutOfSyncException; -061import org.apache.hadoop.hbase.CoordinatedStateManager; -062import org.apache.hadoop.hbase.DoNotRetryIOException; -063import org.apache.hadoop.hbase.HBaseConfiguration; -064import org.apache.hadoop.hbase.HBaseInterfaceAudience; -065import org.apache.hadoop.hbase.HConstants; -066import org.apache.hadoop.hbase.HealthCheckChore; -067import org.apache.hadoop.hbase.MetaTableAccessor; -068import org.apache.hadoop.hbase.NotServingRegionException; -069import org.apache.hadoop.hbase.PleaseHoldException; -070import org.apache.hadoop.hbase.ScheduledChore; -071import org.apache.hadoop.hbase.ServerName; -072import org.apache.hadoop.hbase.Stoppable; -073import org.apache.hadoop.hbase.TableDescriptors; -074import org.apache.hadoop.hbase.TableName; -075import org.apache.hadoop.hbase.YouAreDeadException; -076import org.apache.hadoop.hbase.ZNodeClearer; -077import org.apache.hadoop.hbase.client.ClusterConnection; -078import org.apache.hadoop.hbase.client.Connection; -079import org.apache.hadoop.hbase.client.ConnectionUtils; -080import org.apache.hadoop.hbase.client.RegionInfo; -081import org.apache.hadoop.hbase.client.RegionInfoBuilder; -082import org.apache.hadoop.hbase.client.RpcRetryingCallerFactory; -083import org.apache.hadoop.hbase.client.TableDescriptorBuilder; -084import org.apache.hadoop.hbase.client.locking.EntityLock; -085import org.apache.hadoop.hbase.client.locking.LockServiceClient; -086import org.apache.hadoop.hbase.conf.ConfigurationManager; -087import org.apache.hadoop.hbase.conf.ConfigurationObserver; -088import org.apache.hadoop.hbase.coordination.SplitLogWorkerCoordination; -089import org.apache.hadoop.hbase.coordination.ZkCoordinatedStateManager; -090import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; -091import org.apache.hadoop.hbase.exceptions.RegionMovedException; -092import org.apache.hadoop.hbase.exceptions.RegionOpeningException; -093import org.apache.hadoop.hbase.exceptions.UnknownProtocolException; -094import org.apache.hadoop.hbase.executor.ExecutorService; -095import org.apache.hadoop.hbase.executor.ExecutorType; -096import org.apache.hadoop.hbase.fs.HFileSystem; -097import org.apache.hadoop.hbase.http.InfoServer; -098import org.apache.hadoop.hbase.io.hfile.BlockCache; -099import org.apache.hadoop.hbase.io.hfile.CacheConfig; -100import org.apache.hadoop.hbase.io.hfile.HFile; -101import org.apache.hadoop.hbase.io.util.MemorySizeUtil; -102import org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils; -103import org.apache.hadoop.hbase.ipc.NettyRpcClientConfigHelper; -104import org.apache.hadoop.hbase.ipc.RpcClient; -105import org.apache.hadoop.hbase.ipc.RpcClientFactory; -106import org.apache.hadoop.hbase.ipc.RpcControllerFactory; -107import org.apache.hadoop.hbase.ipc.RpcServer; -108import org.apache.hadoop.hbase.ipc.RpcServerInterface; -109import org.apache.hadoop.hbase.ipc.ServerNotRunningYetException; -110import org.apache.hadoop.hbase.ipc.ServerRpcController; -111import org.apache.hadoop.hbase.log.HBaseMarkers; -112import org.apache.hadoop.hbase.master.HMaster; -113import org.apache.hadoop.hbase.master.LoadBalancer; -114import org.apache.hadoop.hbase.master.RegionState.State;
[20/51] [partial] hbase-site git commit: Published site at 3fe8649b2c9ba1271c25e8f476548907e4c7a90d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8f09a71d/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html index c7d99b2..9d1542c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaHostCostFunction.html @@ -382,1357 +382,1365 @@ 374for (int i = 0; i this.curFunctionCosts.length; i++) { 375 curFunctionCosts[i] = tempFunctionCosts[i]; 376} -377LOG.info("start StochasticLoadBalancer.balancer, initCost=" + currentCost + ", functionCost=" -378+ functionCost()); +377double initCost = currentCost; +378double newCost = currentCost; 379 -380double initCost = currentCost; -381double newCost = currentCost; -382 -383long computedMaxSteps; -384if (runMaxSteps) { -385 computedMaxSteps = Math.max(this.maxSteps, -386 ((long)cluster.numRegions * (long)this.stepsPerRegion * (long)cluster.numServers)); -387} else { -388 computedMaxSteps = Math.min(this.maxSteps, -389 ((long)cluster.numRegions * (long)this.stepsPerRegion * (long)cluster.numServers)); -390} -391// Perform a stochastic walk to see if we can get a good fit. -392long step; -393 -394for (step = 0; step computedMaxSteps; step++) { -395 Cluster.Action action = nextAction(cluster); -396 -397 if (action.type == Type.NULL) { -398continue; -399 } -400 -401 cluster.doAction(action); -402 updateCostsWithAction(cluster, action); -403 -404 newCost = computeCost(cluster, currentCost); -405 -406 // Should this be kept? -407 if (newCost currentCost) { -408currentCost = newCost; -409 -410// save for JMX -411curOverallCost = currentCost; -412for (int i = 0; i this.curFunctionCosts.length; i++) { -413 curFunctionCosts[i] = tempFunctionCosts[i]; -414} -415 } else { -416// Put things back the way they were before. -417// TODO: undo by remembering old values -418Action undoAction = action.undoAction(); -419cluster.doAction(undoAction); -420updateCostsWithAction(cluster, undoAction); -421 } -422 -423 if (EnvironmentEdgeManager.currentTime() - startTime -424 maxRunningTime) { -425break; -426 } -427} -428long endTime = EnvironmentEdgeManager.currentTime(); -429 -430 metricsBalancer.balanceCluster(endTime - startTime); -431 -432// update costs metrics -433updateStochasticCosts(tableName, curOverallCost, curFunctionCosts); -434if (initCost currentCost) { -435 plans = createRegionPlans(cluster); -436 LOG.info("Finished computing new load balance plan. Computation took {}" + -437" to try {} different iterations. Found a solution that moves " + -438"{} regions; Going from a computed cost of {}" + -439" to a new cost of {}", java.time.Duration.ofMillis(endTime - startTime), -440step, plans.size(), initCost, currentCost); -441 return plans; -442} -443LOG.info("Could not find a better load balance plan. Tried {} different configurations in " + -444 "{}, and did not find anything with a computed cost less than {}", step, -445 java.time.Duration.ofMillis(endTime - startTime), initCost); -446return null; -447 } -448 -449 /** -450 * update costs to JMX -451 */ -452 private void updateStochasticCosts(TableName tableName, Double overall, Double[] subCosts) { -453if (tableName == null) return; -454 -455// check if the metricsBalancer is MetricsStochasticBalancer before casting -456if (metricsBalancer instanceof MetricsStochasticBalancer) { -457 MetricsStochasticBalancer balancer = (MetricsStochasticBalancer) metricsBalancer; -458 // overall cost -459 balancer.updateStochasticCost(tableName.getNameAsString(), -460"Overall", "Overall cost", overall); -461 -462 // each cost function -463 for (int i = 0; i costFunctions.length; i++) { -464CostFunction costFunction = costFunctions[i]; -465String costFunctionName = costFunction.getClass().getSimpleName(); -466Double costPercent = (overall == 0) ? 0 : (subCosts[i] / overall); -467// TODO: cost function may need a specific description -468 balancer.updateStochasticCost(tableName.getNameAsString(), costFunctionName, -469 "The percent of " + costFunctionName, costPercent);
[20/51] [partial] hbase-site git commit: Published site at 7adf590106826b9e4432cfeee06acdc0ccff8c6e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/425db230/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html index 57f4692..73e1692 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html +++ b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html @@ -4301,10 +4301,14 @@ static HBaseClassTestRule -TestWALProcedureStore.CLASS_RULE +TestWALProcedureTree.CLASS_RULE static HBaseClassTestRule +TestWALProcedureStore.CLASS_RULE + + +static HBaseClassTestRule TestForceUpdateProcedure.CLASS_RULE http://git-wip-us.apache.org/repos/asf/hbase-site/blob/425db230/testdevapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html b/testdevapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html index ec56445..4837d80 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html +++ b/testdevapidocs/org/apache/hadoop/hbase/io/hfile/package-tree.html @@ -158,8 +158,8 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite.CacheOnWriteType org.apache.hadoop.hbase.io.hfile.TagUsage +org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite.CacheOnWriteType http://git-wip-us.apache.org/repos/asf/hbase-site/blob/425db230/testdevapidocs/org/apache/hadoop/hbase/package-tree.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/package-tree.html b/testdevapidocs/org/apache/hadoop/hbase/package-tree.html index 576f2d6..58e6363 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/package-tree.html +++ b/testdevapidocs/org/apache/hadoop/hbase/package-tree.html @@ -580,12 +580,12 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) org.apache.hadoop.hbase.HBaseClusterManager.CommandProvider.Operation -org.apache.hadoop.hbase.PerformanceEvaluation.Counter -org.apache.hadoop.hbase.IntegrationTestRegionReplicaPerf.Stat org.apache.hadoop.hbase.ScanPerformanceEvaluation.ScanCounter +org.apache.hadoop.hbase.IntegrationTestRegionReplicaPerf.Stat +org.apache.hadoop.hbase.ResourceChecker.Phase org.apache.hadoop.hbase.RESTApiClusterManager.RoleCommand org.apache.hadoop.hbase.IntegrationTestDDLMasterFailover.ACTION -org.apache.hadoop.hbase.ResourceChecker.Phase +org.apache.hadoop.hbase.PerformanceEvaluation.Counter org.apache.hadoop.hbase.ClusterManager.ServiceType org.apache.hadoop.hbase.RESTApiClusterManager.Service http://git-wip-us.apache.org/repos/asf/hbase-site/blob/425db230/testdevapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html b/testdevapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html index 42f874a..e84bc13 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html +++ b/testdevapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html @@ -219,10 +219,10 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.procedure2.TestStateMachineProcedure.TestSMProcedureState org.apache.hadoop.hbase.procedure2.TestProcedureBypass.StuckStateMachineState -org.apache.hadoop.hbase.procedure2.TestProcedureRecovery.TestStateMachineProcedure.State org.apache.hadoop.hbase.procedure2.TestYieldProcedures.TestStateMachineProcedure.State
[20/51] [partial] hbase-site git commit: Published site at 5fbb227deb365fe812d433fe39b85ac4b0ddee20.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html index 1db4dc7..49e372e 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html @@ -1215,25 +1215,30 @@ RSProcedureDispatcher.RegionOpenOperation.buildRegionOpenInfoRequest(MasterProcedureEnvenv) +private boolean +ReopenTableRegionsProcedure.canSchedule(MasterProcedureEnvenv, + HRegionLocationloc) + + protected static void AbstractStateMachineTableProcedure.checkOnline(MasterProcedureEnvenv, RegionInfori) Check region is online. - + protected void AbstractStateMachineRegionProcedure.checkTableModifiable(MasterProcedureEnvenv) Check whether a table is modifiable - exists and either offline or online with config set - + protected void AbstractStateMachineTableProcedure.checkTableModifiable(MasterProcedureEnvenv) Check whether a table is modifiable - exists and either offline or online with config set - + private static void DeleteTableProcedure.cleanAnyRemainingRows(MasterProcedureEnvenv, TableNametableName) @@ -1241,26 +1246,26 @@ info:regioninfo column was empty because of some write error. - + protected void TruncateTableProcedure.completionCleanup(MasterProcedureEnvenv) - + protected void ModifyTableProcedure.completionCleanup(MasterProcedureEnvenv) - + protected void InitMetaProcedure.completionCleanup(MasterProcedureEnvenv) - + protected static void CreateNamespaceProcedure.createDirectory(MasterProcedureEnvenv, NamespaceDescriptornsDescriptor) Create the namespace directory - + private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo CloneSnapshotProcedure.createFilesystemLayout(MasterProcedureEnvenv, TableDescriptortableDescriptor, @@ -1268,20 +1273,20 @@ Create regions in file system. - + protected static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo CreateTableProcedure.createFsLayout(MasterProcedureEnvenv, TableDescriptortableDescriptor, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfonewRegions) - + protected static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo CreateTableProcedure.createFsLayout(MasterProcedureEnvenv, TableDescriptortableDescriptor, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfonewRegions, CreateTableProcedure.CreateHdfsRegionshdfsRegionHandler) - + private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo CloneSnapshotProcedure.createFsLayout(MasterProcedureEnvenv, TableDescriptortableDescriptor, @@ -1290,19 +1295,19 @@ Create region layout in file system. - + https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo CreateTableProcedure.CreateHdfsRegions.createHdfsRegions(MasterProcedureEnvenv, org.apache.hadoop.fs.PathtableRootDir, TableNametableName, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfonewRegions) - + protected static void DeleteTableProcedure.deleteAssignmentState(MasterProcedureEnvenv, TableNametableName) - + static void MasterDDLOperationHelper.deleteColumnFamilyFromFileSystem(MasterProcedureEnvenv, TableNametableName, @@ -1312,14 +1317,14 @@ Remove the column family from the file system - + protected static void DeleteNamespaceProcedure.deleteDirectory(MasterProcedureEnvenv, https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringnamespaceName) Delete the namespace directories from the file system - + private void ModifyTableProcedure.deleteFromFs(MasterProcedureEnvenv, TableDescriptoroldTableDescriptor, @@ -1327,27 +1332,27 @@ Removes from hdfs the families that
[20/51] [partial] hbase-site git commit: Published site at 821e4d7de2d576189f4288d1c2acf9e9a9471f5c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/323b17d9/testdevapidocs/org/apache/hadoop/hbase/client/TestRestoreSnapshotFromClientGetCompactionState.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/client/TestRestoreSnapshotFromClientGetCompactionState.html b/testdevapidocs/org/apache/hadoop/hbase/client/TestRestoreSnapshotFromClientGetCompactionState.html new file mode 100644 index 000..078c52a --- /dev/null +++ b/testdevapidocs/org/apache/hadoop/hbase/client/TestRestoreSnapshotFromClientGetCompactionState.html @@ -0,0 +1,369 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +TestRestoreSnapshotFromClientGetCompactionState (Apache HBase 3.0.0-SNAPSHOT Test API) + + + + + +var methods = {"i0":10,"i1":9}; +var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Field| +Constr| +Method + + +Detail: +Field| +Constr| +Method + + + + + + + + +org.apache.hadoop.hbase.client +Class TestRestoreSnapshotFromClientGetCompactionState + + + +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.client.RestoreSnapshotFromClientTestBase + + +org.apache.hadoop.hbase.client.RestoreSnapshotFromClientGetCompactionStateTestBase + + +org.apache.hadoop.hbase.client.TestRestoreSnapshotFromClientGetCompactionState + + + + + + + + + + + + +public class TestRestoreSnapshotFromClientGetCompactionState +extends RestoreSnapshotFromClientGetCompactionStateTestBase + + + + + + + + + + + +Field Summary + +Fields + +Modifier and Type +Field and Description + + +static HBaseClassTestRule +CLASS_RULE + + +int +numReplicas + + + + + + +Fields inherited from classorg.apache.hadoop.hbase.client.RestoreSnapshotFromClientTestBase +admin, emptySnapshot, FAMILY, name, snapshot0Rows, snapshot1Rows, snapshotName0, snapshotName1, snapshotName2, tableName, TEST_FAMILY2, TEST_UTIL + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +TestRestoreSnapshotFromClientGetCompactionState() + + + + + + + + + +Method Summary + +All MethodsStatic MethodsInstance MethodsConcrete Methods + +Modifier and Type +Method and Description + + +protected int +getNumReplicas() + + +static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttps://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object[] +params() + + + + + + +Methods inherited from classorg.apache.hadoop.hbase.client.RestoreSnapshotFromClientGetCompactionStateTestBase +testGetCompactionStateAfterRestoringSnapshot + + + + + +Methods inherited from classorg.apache.hadoop.hbase.client.RestoreSnapshotFromClientTestBase +countRows, createTable, getValidMethodName, setup, setupCluster, setupConf, splitRe gion, tearDown, tearDownAfterClass, verifyRowCount + + + + + +Methods inherited from classjava.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--; title="class or interface in java.lang">clone, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-; title="class or interface in java.lang">equals, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--; title="class or interface in java.lang">finalize, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--; title="class or interface in java.lang">getClass, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--; title="class or interface in java.lang">hashCode, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--; title="class or interface in java.lang">notify,
[20/51] [partial] hbase-site git commit: Published site at fa5fa6ecdd071b72b58971058ff3ab9d28c3e709.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d1341859/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.ProcedureStoreListener.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.ProcedureStoreListener.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.ProcedureStoreListener.html index ff29160..e4dc134 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.ProcedureStoreListener.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.ProcedureStoreListener.html @@ -42,190 +42,208 @@ 034public interface ProcedureStore { 035 /** 036 * Store listener interface. -037 * The main process should register a listener and respond to the store events. -038 */ -039 public interface ProcedureStoreListener { -040/** -041 * triggered when the store sync is completed. -042 */ -043void postSync(); -044 -045/** -046 * triggered when the store is not able to write out data. -047 * the main process should abort. -048 */ -049void abortProcess(); -050 } -051 -052 /** -053 * An Iterator over a collection of Procedure -054 */ -055 public interface ProcedureIterator { -056/** -057 * Reset the Iterator by seeking to the beginning of the list. -058 */ -059void reset(); -060 -061/** -062 * Returns true if the iterator has more elements. -063 * (In other words, returns true if next() would return a Procedure -064 * rather than throwing an exception.) -065 * @return true if the iterator has more procedures -066 */ -067boolean hasNext(); -068 -069/** -070 * @return true if the iterator next element is a completed procedure. -071 */ -072boolean isNextFinished(); -073 +037 * p/ +038 * The main process should register a listener and respond to the store events. +039 */ +040 public interface ProcedureStoreListener { +041 +042/** +043 * triggered when the store sync is completed. +044 */ +045default void postSync() { +046} +047 +048/** +049 * triggered when the store is not able to write out data. the main process should abort. +050 */ +051default void abortProcess() { +052} +053 +054/** +055 * Suggest that the upper layer should update the state of some procedures. Ignore this call +056 * will not effect correctness but performance. +057 * p/ +058 * For a WAL based ProcedureStore implementation, if all the procedures stored in a WAL file +059 * have been deleted, or updated later in another WAL file, then we can delete the WAL file. If +060 * there are old procedures in a WAL file which are never deleted or updated, then we can not +061 * delete the WAL file and this will cause we hold lots of WAL file and slow down the master +062 * restarts. So here we introduce this method to tell the upper layer that please update the +063 * states of these procedures so that we can delete the old WAL file. +064 * @param procIds the id for the procedures +065 */ +066default void forceUpdate(long[] procIds) { +067} +068 } +069 +070 /** +071 * An Iterator over a collection of Procedure +072 */ +073 public interface ProcedureIterator { 074/** -075 * Skip the next procedure +075 * Reset the Iterator by seeking to the beginning of the list. 076 */ -077void skipNext(); +077void reset(); 078 079/** -080 * Returns the next procedure in the iteration. -081 * @throws IOException if there was an error fetching/deserializing the procedure -082 * @return the next procedure in the iteration. -083 */ -084@SuppressWarnings("rawtypes") -085Procedure next() throws IOException; -086 } -087 -088 /** -089 * Interface passed to the ProcedureStore.load() method to handle the store-load events. -090 */ -091 public interface ProcedureLoader { +080 * Returns true if the iterator has more elements. +081 * (In other words, returns true if next() would return a Procedure +082 * rather than throwing an exception.) +083 * @return true if the iterator has more procedures +084 */ +085boolean hasNext(); +086 +087/** +088 * @return true if the iterator next element is a completed procedure. +089 */ +090boolean isNextFinished(); +091 092/** -093 * Called by ProcedureStore.load() to notify about the maximum proc-id in the store. -094 * @param maxProcId the highest proc-id in the store -095 */ -096void setMaxProcId(long maxProcId); -097 -098/** -099 * Called by the ProcedureStore.load() every time a set of procedures are ready to be executed. -100 * The ProcedureIterator passed to the method, has the procedure sorted in replay-order. -101 *
[20/51] [partial] hbase-site git commit: Published site at 6bc7089f9e0793efc9bdd46a84f5ccd9bc4579ad.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/419d0338/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerMonitor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerMonitor.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerMonitor.html index 43c66a8..061ce80 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerMonitor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.WorkerMonitor.html @@ -23,2136 +23,2142 @@ 015 * See the License for the specific language governing permissions and 016 * limitations under the License. 017 */ -018 -019package org.apache.hadoop.hbase.procedure2; -020 -021import java.io.IOException; -022import java.util.ArrayDeque; -023import java.util.ArrayList; -024import java.util.Arrays; -025import java.util.Collection; -026import java.util.Deque; -027import java.util.HashSet; -028import java.util.Iterator; -029import java.util.List; -030import java.util.Map; -031import java.util.Objects; -032import java.util.Set; -033import java.util.concurrent.ConcurrentHashMap; -034import java.util.concurrent.CopyOnWriteArrayList; -035import java.util.concurrent.TimeUnit; -036import java.util.concurrent.atomic.AtomicBoolean; -037import java.util.concurrent.atomic.AtomicInteger; -038import java.util.concurrent.atomic.AtomicLong; -039import java.util.stream.Collectors; -040import java.util.stream.Stream; -041 -042import org.apache.hadoop.conf.Configuration; -043import org.apache.hadoop.hbase.HConstants; -044import org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException; -045import org.apache.hadoop.hbase.log.HBaseMarkers; -046import org.apache.hadoop.hbase.procedure2.Procedure.LockState; -047import org.apache.hadoop.hbase.procedure2.store.ProcedureStore; -048import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator; -049import org.apache.hadoop.hbase.procedure2.util.StringUtils; -050import org.apache.hadoop.hbase.security.User; -051import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -052import org.apache.hadoop.hbase.util.IdLock; -053import org.apache.hadoop.hbase.util.NonceKey; -054import org.apache.hadoop.hbase.util.Threads; -055import org.apache.yetus.audience.InterfaceAudience; -056import org.slf4j.Logger; -057import org.slf4j.LoggerFactory; -058 -059import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; -060import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; +018package org.apache.hadoop.hbase.procedure2; +019 +020import java.io.IOException; +021import java.util.ArrayDeque; +022import java.util.ArrayList; +023import java.util.Arrays; +024import java.util.Collection; +025import java.util.Deque; +026import java.util.HashSet; +027import java.util.Iterator; +028import java.util.List; +029import java.util.Map; +030import java.util.Objects; +031import java.util.Set; +032import java.util.concurrent.ConcurrentHashMap; +033import java.util.concurrent.CopyOnWriteArrayList; +034import java.util.concurrent.TimeUnit; +035import java.util.concurrent.atomic.AtomicBoolean; +036import java.util.concurrent.atomic.AtomicInteger; +037import java.util.concurrent.atomic.AtomicLong; +038import java.util.stream.Collectors; +039import java.util.stream.Stream; +040import org.apache.hadoop.conf.Configuration; +041import org.apache.hadoop.hbase.HConstants; +042import org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException; +043import org.apache.hadoop.hbase.log.HBaseMarkers; +044import org.apache.hadoop.hbase.procedure2.Procedure.LockState; +045import org.apache.hadoop.hbase.procedure2.store.ProcedureStore; +046import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator; +047import org.apache.hadoop.hbase.procedure2.util.StringUtils; +048import org.apache.hadoop.hbase.security.User; +049import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; +050import org.apache.hadoop.hbase.util.IdLock; +051import org.apache.hadoop.hbase.util.NonceKey; +052import org.apache.hadoop.hbase.util.Threads; +053import org.apache.yetus.audience.InterfaceAudience; +054import org.slf4j.Logger; +055import org.slf4j.LoggerFactory; +056 +057import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; +058import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; +059 +060import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState; 061 -062import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState; -063 -064/** -065 * Thread Pool that executes the submitted procedures. -066 * The executor has a ProcedureStore associated. -067 * Each operation is logged and on restart the pending procedures are resumed. -068 * -069 * Unless the
[20/51] [partial] hbase-site git commit: Published site at d7e08317d2f214e4cca7b67578aba0ed7a567d54.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html index 976894f..721035e 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.InitializationMonitor.html @@ -3020,926 +3020,927 @@ 3012} 3013 } 3014 -3015 void checkServiceStarted() throws ServerNotRunningYetException { -3016if (!serviceStarted) { -3017 throw new ServerNotRunningYetException("Server is not running yet"); -3018} -3019 } -3020 -3021 public static class MasterStoppedException extends DoNotRetryIOException { -3022MasterStoppedException() { -3023 super(); -3024} -3025 } -3026 -3027 void checkInitialized() throws PleaseHoldException, ServerNotRunningYetException, -3028 MasterNotRunningException, MasterStoppedException { -3029checkServiceStarted(); -3030if (!isInitialized()) { -3031 throw new PleaseHoldException("Master is initializing"); -3032} -3033if (isStopped()) { -3034 throw new MasterStoppedException(); -3035} -3036 } -3037 -3038 /** -3039 * Report whether this master is currently the active master or not. -3040 * If not active master, we are parked on ZK waiting to become active. -3041 * -3042 * This method is used for testing. -3043 * -3044 * @return true if active master, false if not. -3045 */ -3046 @Override -3047 public boolean isActiveMaster() { -3048return activeMaster; -3049 } -3050 -3051 /** -3052 * Report whether this master has completed with its initialization and is -3053 * ready. If ready, the master is also the active master. A standby master -3054 * is never ready. -3055 * -3056 * This method is used for testing. -3057 * -3058 * @return true if master is ready to go, false if not. -3059 */ -3060 @Override -3061 public boolean isInitialized() { -3062return initialized.isReady(); -3063 } -3064 -3065 /** -3066 * Report whether this master is in maintenance mode. -3067 * -3068 * @return true if master is in maintenanceMode -3069 */ -3070 @Override -3071 public boolean isInMaintenanceMode() throws IOException { -3072if (!isInitialized()) { -3073 throw new PleaseHoldException("Master is initializing"); -3074} -3075return maintenanceModeTracker.isInMaintenanceMode(); -3076 } -3077 -3078 @VisibleForTesting -3079 public void setInitialized(boolean isInitialized) { -3080 procedureExecutor.getEnvironment().setEventReady(initialized, isInitialized); -3081 } -3082 -3083 @Override -3084 public ProcedureEvent? getInitializedEvent() { -3085return initialized; -3086 } -3087 -3088 /** -3089 * Compute the average load across all region servers. -3090 * Currently, this uses a very naive computation - just uses the number of -3091 * regions being served, ignoring stats about number of requests. -3092 * @return the average load -3093 */ -3094 public double getAverageLoad() { -3095if (this.assignmentManager == null) { -3096 return 0; -3097} -3098 -3099RegionStates regionStates = this.assignmentManager.getRegionStates(); -3100if (regionStates == null) { -3101 return 0; -3102} -3103return regionStates.getAverageLoad(); -3104 } -3105 -3106 /* -3107 * @return the count of region split plans executed -3108 */ -3109 public long getSplitPlanCount() { -3110return splitPlanCount; -3111 } -3112 -3113 /* -3114 * @return the count of region merge plans executed -3115 */ -3116 public long getMergePlanCount() { -3117return mergePlanCount; -3118 } -3119 -3120 @Override -3121 public boolean registerService(Service instance) { -3122/* -3123 * No stacking of instances is allowed for a single service name -3124 */ -3125Descriptors.ServiceDescriptor serviceDesc = instance.getDescriptorForType(); -3126String serviceName = CoprocessorRpcUtils.getServiceName(serviceDesc); -3127if (coprocessorServiceHandlers.containsKey(serviceName)) { -3128 LOG.error("Coprocessor service "+serviceName+ -3129 " already registered, rejecting request from "+instance -3130 ); -3131 return false; -3132} -3133 -3134 coprocessorServiceHandlers.put(serviceName, instance); -3135if (LOG.isDebugEnabled()) { -3136 LOG.debug("Registered master coprocessor service: service="+serviceName); -3137} -3138return true; -3139 } -3140 -3141 /** -3142 * Utility for constructing an instance of the passed HMaster class. -3143 * @param masterClass -3144 * @return HMaster instance. -3145 */ -3146
[20/51] [partial] hbase-site git commit: Published site at 8eaaa63114a64bcaeaf0ed9bdd88615ee22255c1.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f6f9d4f3/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html index 0cf012a..976894f 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html @@ -63,3884 +63,3883 @@ 055import javax.servlet.http.HttpServlet; 056import javax.servlet.http.HttpServletRequest; 057import javax.servlet.http.HttpServletResponse; -058 -059import org.apache.commons.lang3.StringUtils; -060import org.apache.hadoop.conf.Configuration; -061import org.apache.hadoop.fs.Path; -062import org.apache.hadoop.hbase.ChoreService; -063import org.apache.hadoop.hbase.ClusterId; -064import org.apache.hadoop.hbase.ClusterMetrics; -065import org.apache.hadoop.hbase.ClusterMetrics.Option; -066import org.apache.hadoop.hbase.ClusterMetricsBuilder; -067import org.apache.hadoop.hbase.CompoundConfiguration; -068import org.apache.hadoop.hbase.DoNotRetryIOException; -069import org.apache.hadoop.hbase.HBaseIOException; -070import org.apache.hadoop.hbase.HBaseInterfaceAudience; -071import org.apache.hadoop.hbase.HConstants; -072import org.apache.hadoop.hbase.InvalidFamilyOperationException; -073import org.apache.hadoop.hbase.MasterNotRunningException; -074import org.apache.hadoop.hbase.MetaTableAccessor; -075import org.apache.hadoop.hbase.NamespaceDescriptor; -076import org.apache.hadoop.hbase.PleaseHoldException; -077import org.apache.hadoop.hbase.ReplicationPeerNotFoundException; -078import org.apache.hadoop.hbase.ServerName; -079import org.apache.hadoop.hbase.TableDescriptors; -080import org.apache.hadoop.hbase.TableName; -081import org.apache.hadoop.hbase.TableNotDisabledException; -082import org.apache.hadoop.hbase.TableNotFoundException; -083import org.apache.hadoop.hbase.UnknownRegionException; -084import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor; -085import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder; -086import org.apache.hadoop.hbase.client.MasterSwitchType; -087import org.apache.hadoop.hbase.client.RegionInfo; -088import org.apache.hadoop.hbase.client.RegionInfoBuilder; -089import org.apache.hadoop.hbase.client.Result; -090import org.apache.hadoop.hbase.client.TableDescriptor; -091import org.apache.hadoop.hbase.client.TableDescriptorBuilder; -092import org.apache.hadoop.hbase.client.TableState; -093import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; -094import org.apache.hadoop.hbase.exceptions.DeserializationException; -095import org.apache.hadoop.hbase.exceptions.MergeRegionException; -096import org.apache.hadoop.hbase.executor.ExecutorType; -097import org.apache.hadoop.hbase.favored.FavoredNodesManager; -098import org.apache.hadoop.hbase.favored.FavoredNodesPromoter; -099import org.apache.hadoop.hbase.http.InfoServer; -100import org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils; -101import org.apache.hadoop.hbase.ipc.RpcServer; -102import org.apache.hadoop.hbase.ipc.ServerNotRunningYetException; -103import org.apache.hadoop.hbase.log.HBaseMarkers; -104import org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode; -105import org.apache.hadoop.hbase.master.assignment.AssignProcedure; -106import org.apache.hadoop.hbase.master.assignment.AssignmentManager; -107import org.apache.hadoop.hbase.master.assignment.MergeTableRegionsProcedure; -108import org.apache.hadoop.hbase.master.assignment.MoveRegionProcedure; -109import org.apache.hadoop.hbase.master.assignment.RegionStateNode; -110import org.apache.hadoop.hbase.master.assignment.RegionStates; -111import org.apache.hadoop.hbase.master.assignment.TransitRegionStateProcedure; -112import org.apache.hadoop.hbase.master.assignment.UnassignProcedure; -113import org.apache.hadoop.hbase.master.balancer.BalancerChore; -114import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer; -115import org.apache.hadoop.hbase.master.balancer.ClusterStatusChore; -116import org.apache.hadoop.hbase.master.balancer.LoadBalancerFactory; -117import org.apache.hadoop.hbase.master.cleaner.CleanerChore; -118import org.apache.hadoop.hbase.master.cleaner.HFileCleaner; -119import org.apache.hadoop.hbase.master.cleaner.LogCleaner; -120import org.apache.hadoop.hbase.master.cleaner.ReplicationBarrierCleaner; -121import org.apache.hadoop.hbase.master.locking.LockManager; -122import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan; -123import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan.PlanType; -124import org.apache.hadoop.hbase.master.normalizer.RegionNormalizer; -125import org.apache.hadoop.hbase.master.normalizer.RegionNormalizerChore; -126import org.apache.hadoop.hbase.master.normalizer.RegionNormalizerFactory;
[20/51] [partial] hbase-site git commit: Published site at cd161d976ef47b84e904f2d54bac65d2f3417c2a.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fa1bebf8/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TruncateTableFuture.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TruncateTableFuture.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TruncateTableFuture.html index a5789e0..93a57cb 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TruncateTableFuture.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TruncateTableFuture.html @@ -238,4120 +238,4119 @@ 230 * @see Admin 231 */ 232@InterfaceAudience.Private -233@InterfaceStability.Evolving -234public class HBaseAdmin implements Admin { -235 private static final Logger LOG = LoggerFactory.getLogger(HBaseAdmin.class); -236 -237 private ClusterConnection connection; -238 -239 private final Configuration conf; -240 private final long pause; -241 private final int numRetries; -242 private final int syncWaitTimeout; -243 private boolean aborted; -244 private int operationTimeout; -245 private int rpcTimeout; -246 -247 private RpcRetryingCallerFactory rpcCallerFactory; -248 private RpcControllerFactory rpcControllerFactory; -249 -250 private NonceGenerator ng; -251 -252 @Override -253 public int getOperationTimeout() { -254return operationTimeout; -255 } -256 -257 HBaseAdmin(ClusterConnection connection) throws IOException { -258this.conf = connection.getConfiguration(); -259this.connection = connection; -260 -261// TODO: receive ConnectionConfiguration here rather than re-parsing these configs every time. -262this.pause = this.conf.getLong(HConstants.HBASE_CLIENT_PAUSE, -263 HConstants.DEFAULT_HBASE_CLIENT_PAUSE); -264this.numRetries = this.conf.getInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, -265 HConstants.DEFAULT_HBASE_CLIENT_RETRIES_NUMBER); -266this.operationTimeout = this.conf.getInt(HConstants.HBASE_CLIENT_OPERATION_TIMEOUT, -267 HConstants.DEFAULT_HBASE_CLIENT_OPERATION_TIMEOUT); -268this.rpcTimeout = this.conf.getInt(HConstants.HBASE_RPC_TIMEOUT_KEY, -269 HConstants.DEFAULT_HBASE_RPC_TIMEOUT); -270this.syncWaitTimeout = this.conf.getInt( -271 "hbase.client.sync.wait.timeout.msec", 10 * 6); // 10min -272 -273this.rpcCallerFactory = connection.getRpcRetryingCallerFactory(); -274this.rpcControllerFactory = connection.getRpcControllerFactory(); -275 -276this.ng = this.connection.getNonceGenerator(); -277 } -278 -279 @Override -280 public void abort(String why, Throwable e) { -281// Currently does nothing but throw the passed message and exception -282this.aborted = true; -283throw new RuntimeException(why, e); -284 } -285 -286 @Override -287 public boolean isAborted() { -288return this.aborted; -289 } -290 -291 @Override -292 public boolean abortProcedure(final long procId, final boolean mayInterruptIfRunning) -293 throws IOException { -294return get(abortProcedureAsync(procId, mayInterruptIfRunning), this.syncWaitTimeout, -295 TimeUnit.MILLISECONDS); -296 } -297 -298 @Override -299 public FutureBoolean abortProcedureAsync(final long procId, final boolean mayInterruptIfRunning) -300 throws IOException { -301Boolean abortProcResponse = -302executeCallable(new MasterCallableAbortProcedureResponse(getConnection(), -303getRpcControllerFactory()) { -304 @Override -305 protected AbortProcedureResponse rpcCall() throws Exception { -306AbortProcedureRequest abortProcRequest = -307 AbortProcedureRequest.newBuilder().setProcId(procId).build(); -308return master.abortProcedure(getRpcController(), abortProcRequest); -309 } -310}).getIsProcedureAborted(); -311return new AbortProcedureFuture(this, procId, abortProcResponse); -312 } -313 -314 @Override -315 public ListTableDescriptor listTableDescriptors() throws IOException { -316return listTableDescriptors((Pattern)null, false); -317 } -318 -319 @Override -320 public ListTableDescriptor listTableDescriptors(Pattern pattern) throws IOException { -321return listTableDescriptors(pattern, false); -322 } -323 -324 @Override -325 public ListTableDescriptor listTableDescriptors(Pattern pattern, boolean includeSysTables) -326 throws IOException { -327return executeCallable(new MasterCallableListTableDescriptor(getConnection(), -328getRpcControllerFactory()) { -329 @Override -330 protected ListTableDescriptor rpcCall() throws Exception { -331GetTableDescriptorsRequest req = -332 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables); -333return ProtobufUtil.toTableDescriptorList(master.getTableDescriptors(getRpcController(), -334req)); -335 } -336}); -337 } -338
[20/51] [partial] hbase-site git commit: Published site at c6a65ba63fce85ac7c4b62b96ef2bbe6c35d2f00.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/293abb17/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html index fe4e081..eecf20f 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/AssignmentManager.RegionInTransitionChore.html @@ -44,1858 +44,1838 @@ 036import org.apache.hadoop.hbase.HBaseIOException; 037import org.apache.hadoop.hbase.HConstants; 038import org.apache.hadoop.hbase.PleaseHoldException; -039import org.apache.hadoop.hbase.RegionException; -040import org.apache.hadoop.hbase.ServerName; -041import org.apache.hadoop.hbase.TableName; -042import org.apache.hadoop.hbase.UnknownRegionException; -043import org.apache.hadoop.hbase.YouAreDeadException; -044import org.apache.hadoop.hbase.client.DoNotRetryRegionException; -045import org.apache.hadoop.hbase.client.RegionInfo; -046import org.apache.hadoop.hbase.client.RegionInfoBuilder; -047import org.apache.hadoop.hbase.client.Result; -048import org.apache.hadoop.hbase.client.TableState; -049import org.apache.hadoop.hbase.exceptions.UnexpectedStateException; -050import org.apache.hadoop.hbase.favored.FavoredNodesManager; -051import org.apache.hadoop.hbase.favored.FavoredNodesPromoter; -052import org.apache.hadoop.hbase.master.LoadBalancer; -053import org.apache.hadoop.hbase.master.MasterServices; -054import org.apache.hadoop.hbase.master.MetricsAssignmentManager; -055import org.apache.hadoop.hbase.master.NoSuchProcedureException; -056import org.apache.hadoop.hbase.master.RegionPlan; -057import org.apache.hadoop.hbase.master.RegionState; -058import org.apache.hadoop.hbase.master.RegionState.State; -059import org.apache.hadoop.hbase.master.ServerListener; -060import org.apache.hadoop.hbase.master.TableStateManager; -061import org.apache.hadoop.hbase.master.balancer.FavoredStochasticBalancer; -062import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv; -063import org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler; -064import org.apache.hadoop.hbase.master.procedure.ProcedureSyncWait; -065import org.apache.hadoop.hbase.master.procedure.ServerCrashProcedure; -066import org.apache.hadoop.hbase.procedure2.Procedure; -067import org.apache.hadoop.hbase.procedure2.ProcedureEvent; -068import org.apache.hadoop.hbase.procedure2.ProcedureExecutor; -069import org.apache.hadoop.hbase.procedure2.ProcedureInMemoryChore; -070import org.apache.hadoop.hbase.procedure2.util.StringUtils; -071import org.apache.hadoop.hbase.regionserver.SequenceId; -072import org.apache.hadoop.hbase.util.Bytes; -073import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -074import org.apache.hadoop.hbase.util.HasThread; -075import org.apache.hadoop.hbase.util.Pair; -076import org.apache.hadoop.hbase.util.Threads; -077import org.apache.hadoop.hbase.util.VersionInfo; -078import org.apache.hadoop.hbase.zookeeper.MetaTableLocator; -079import org.apache.hadoop.hbase.zookeeper.ZKWatcher; -080import org.apache.yetus.audience.InterfaceAudience; -081import org.apache.zookeeper.KeeperException; -082import org.slf4j.Logger; -083import org.slf4j.LoggerFactory; +039import org.apache.hadoop.hbase.ServerName; +040import org.apache.hadoop.hbase.TableName; +041import org.apache.hadoop.hbase.UnknownRegionException; +042import org.apache.hadoop.hbase.YouAreDeadException; +043import org.apache.hadoop.hbase.client.DoNotRetryRegionException; +044import org.apache.hadoop.hbase.client.RegionInfo; +045import org.apache.hadoop.hbase.client.RegionInfoBuilder; +046import org.apache.hadoop.hbase.client.Result; +047import org.apache.hadoop.hbase.client.TableState; +048import org.apache.hadoop.hbase.exceptions.UnexpectedStateException; +049import org.apache.hadoop.hbase.favored.FavoredNodesManager; +050import org.apache.hadoop.hbase.favored.FavoredNodesPromoter; +051import org.apache.hadoop.hbase.master.LoadBalancer; +052import org.apache.hadoop.hbase.master.MasterServices; +053import org.apache.hadoop.hbase.master.MetricsAssignmentManager; +054import org.apache.hadoop.hbase.master.RegionPlan; +055import org.apache.hadoop.hbase.master.RegionState; +056import org.apache.hadoop.hbase.master.RegionState.State; +057import org.apache.hadoop.hbase.master.ServerListener; +058import org.apache.hadoop.hbase.master.TableStateManager; +059import org.apache.hadoop.hbase.master.balancer.FavoredStochasticBalancer; +060import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv; +061import
[20/51] [partial] hbase-site git commit: Published site at 7c1fad4992a169a35b4457e6f4afcb30d04406e9.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/74f60271/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTableWrapperAggregateImpl.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTableWrapperAggregateImpl.html b/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTableWrapperAggregateImpl.html index 74ae9d2..bba6ed2 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTableWrapperAggregateImpl.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTableWrapperAggregateImpl.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -220,44 +220,94 @@ implements long +getAvgRegionSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) +Get the average region size for this table + + + +long +getAvgStoreFileAge(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) + + +long +getCpRequestCount(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) + + +long getCpRequestsCount(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) Get the number of CoprocessorService requests that have been issued against this table - + +long +getFilteredReadRequestCount(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) +Get the total number of filtered read requests that have been issued against this table + + + long -getMemStoresSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) +getMaxStoreFileAge(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) + + +long +getMemStoreSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) Get the memory store size against this table - + +long +getMinStoreFileAge(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) + + +long +getNumReferenceFiles(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) + + +long +getNumRegions(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) +Get the number of regions hosted on for this table + + + +long +getNumStoreFiles(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) +Get the number of store files hosted for this table + + + +long +getNumStores(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) +Get the number of stores hosted on for this table + + + long -getReadRequestsCount(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) +getReadRequestCount(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) Get the number of read requests that have been issued against this table - + long -getStoreFilesSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) +getStoreFileSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) Get the store file size against this table - + long getTableSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) Get the table region size against this table - + long getTotalRequestsCount(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) -Get the total number of requests that have been issued against this table +Get the total number of requests that have been issued for this
[20/51] [partial] hbase-site git commit: Published site at 3afe9fb7e6ebfa71187cbe131558a83fae61cecd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/downloads.html -- diff --git a/downloads.html b/downloads.html index a8552e1..83b6734 100644 --- a/downloads.html +++ b/downloads.html @@ -7,7 +7,7 @@ - + Apache HBase Apache HBase Downloads @@ -423,7 +423,7 @@ under the License. --> https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-08-27 + Last Published: 2018-08-28 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/export_control.html -- diff --git a/export_control.html b/export_control.html index 793f8ef..aed354c 100644 --- a/export_control.html +++ b/export_control.html @@ -7,7 +7,7 @@ - + Apache HBase Export Control @@ -331,7 +331,7 @@ for more details. https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-08-27 + Last Published: 2018-08-28 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/index.html -- diff --git a/index.html b/index.html index 1bf2722..ef12102 100644 --- a/index.html +++ b/index.html @@ -7,7 +7,7 @@ - + Apache HBase Apache HBase⢠Home @@ -411,7 +411,7 @@ Apache HBase is an open-source, distributed, versioned, non-relational database https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-08-27 + Last Published: 2018-08-28 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/integration.html -- diff --git a/integration.html b/integration.html index 08000a6..6904d39 100644 --- a/integration.html +++ b/integration.html @@ -7,7 +7,7 @@ - + Apache HBase CI Management @@ -291,7 +291,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-08-27 + Last Published: 2018-08-28 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/issue-tracking.html -- diff --git a/issue-tracking.html b/issue-tracking.html index 62b64ee..75b584b 100644 --- a/issue-tracking.html +++ b/issue-tracking.html @@ -7,7 +7,7 @@ - + Apache HBase Issue Management @@ -288,7 +288,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-08-27 + Last Published: 2018-08-28 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/license.html -- diff --git a/license.html b/license.html index 7a4f11f..e1199d3 100644 --- a/license.html +++ b/license.html @@ -7,7 +7,7 @@ - + Apache HBase Project Licenses @@ -491,7 +491,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-08-27 + Last Published: 2018-08-28 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/mail-lists.html -- diff --git a/mail-lists.html b/mail-lists.html index a825b70..845ab02 100644 --- a/mail-lists.html +++ b/mail-lists.html @@ -7,7 +7,7 @@ - + Apache HBase Project Mailing Lists @@ -341,7 +341,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-08-27 + Last Published: 2018-08-28 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/metrics.html -- diff --git a/metrics.html b/metrics.html index 6a00fab..d9a7ce0 100644 --- a/metrics.html +++ b/metrics.html
[20/51] [partial] hbase-site git commit: Published site at a452487a9b82bfd33bc10683c3f8b8ae74d58883.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cf79db0/devapidocs/org/apache/hadoop/hbase/filter/ColumnValueFilter.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/filter/ColumnValueFilter.html b/devapidocs/org/apache/hadoop/hbase/filter/ColumnValueFilter.html index dab230a..2ccd245 100644 --- a/devapidocs/org/apache/hadoop/hbase/filter/ColumnValueFilter.html +++ b/devapidocs/org/apache/hadoop/hbase/filter/ColumnValueFilter.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":9,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":9,"i12":10,"i13":10,"i14":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":9,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":9,"i14":10,"i15":10,"i16":10}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -120,7 +120,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Public -public class ColumnValueFilter +public class ColumnValueFilter extends FilterBase Different from SingleColumnValueFilter which returns an entire row when specified condition is matched, ColumnValueFilter return the matched cell only. @@ -254,59 +254,67 @@ extends +boolean +equals(https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Objectobj) + + Filter.ReturnCode filterCell(Cellc) A way to filter based on the column family, column qualifier and/or the column value. - + boolean filterRowKey(Cellcell) Filters a row based on the row key. - + ByteArrayComparable getComparator() - + CompareOperator getCompareOperator() - + byte[] getFamily() - + byte[] getQualifier() - + +int +hashCode() + + boolean isFamilyEssential(byte[]name) By default, we require all scan's column families to be present. - + static ColumnValueFilter parseFrom(byte[]pbBytes) Parse protobuf bytes to a ColumnValueFilter - + void reset() Filters that are purely stateless and do nothing in their reset() methods can inherit this null/empty implementation. - + byte[] toByteArray() Return length 0 byte array for Filters that don't require special serialization - + https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String toString() Return filter's info for debugging and logging purpose. @@ -332,7 +340,7 @@ extends Methods inherited from classjava.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object -https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--; title="class or interface in java.lang">clone, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-; title="class or interface in java.lang">equals, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--; title="class or interface in java.lang">finalize, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--; title="class or interface in java.lang">getClass, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--; title="class or interface in java.lang">hashCode, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--; title="class or interface in java.lang">notify, https://docs.oracle.com/javase/8/docs/api/ja va/lang/Object.html?is-external=true#notifyAll--" title="class or interface in java.lang">notifyAll, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait--; title="class or interface in java.lang">wait, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-; title="class or interface in java.lang">wait, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-int-; title="class or interface in java.lang">wait +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--; title="class or interface in java.lang">clone, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--; title="class or interface in java.lang">finalize, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--; title="class or interface in java.lang">getClass, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--; title="class or interface in java.lang">notify, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notifyAll--; title="class or interface in java.lang">notifyAll,
[20/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html index db8431b..a8cb7c4 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html @@ -885,7766 +885,7797 @@ 877 * @return What the next sequence (edit) id should be. 878 * @throws IOException e 879 */ -880 private long initialize(final CancelableProgressable reporter) throws IOException { -881 -882//Refuse to open the region if there is no column family in the table -883if (htableDescriptor.getColumnFamilyCount() == 0) { -884 throw new DoNotRetryIOException("Table " + htableDescriptor.getTableName().getNameAsString()+ -885 " should have at least one column family."); -886} -887 -888MonitoredTask status = TaskMonitor.get().createStatus("Initializing region " + this); -889long nextSeqId = -1; -890try { -891 nextSeqId = initializeRegionInternals(reporter, status); -892 return nextSeqId; -893} finally { -894 // nextSeqid will be -1 if the initialization fails. -895 // At least it will be 0 otherwise. -896 if (nextSeqId == -1) { -897status.abort("Exception during region " + getRegionInfo().getRegionNameAsString() + -898 " initialization."); -899 } -900} -901 } -902 -903 private long initializeRegionInternals(final CancelableProgressable reporter, -904 final MonitoredTask status) throws IOException { -905if (coprocessorHost != null) { -906 status.setStatus("Running coprocessor pre-open hook"); -907 coprocessorHost.preOpen(); -908} -909 -910// Write HRI to a file in case we need to recover hbase:meta -911// Only the primary replica should write .regioninfo -912if (this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) { -913 status.setStatus("Writing region info on filesystem"); -914 fs.checkRegionInfoOnFilesystem(); -915} -916 -917// Initialize all the HStores -918status.setStatus("Initializing all the Stores"); -919long maxSeqId = initializeStores(reporter, status); -920this.mvcc.advanceTo(maxSeqId); -921if (ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) { -922 CollectionHStore stores = this.stores.values(); -923 try { -924// update the stores that we are replaying -925LOG.debug("replaying wal for " + this.getRegionInfo().getEncodedName()); -926 stores.forEach(HStore::startReplayingFromWAL); -927// Recover any edits if available. -928maxSeqId = Math.max(maxSeqId, -929 replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, status)); -930// Make sure mvcc is up to max. -931this.mvcc.advanceTo(maxSeqId); -932 } finally { -933LOG.debug("stopping wal replay for " + this.getRegionInfo().getEncodedName()); -934// update the stores that we are done replaying -935 stores.forEach(HStore::stopReplayingFromWAL); -936 } -937} -938this.lastReplayedOpenRegionSeqId = maxSeqId; -939 -940 this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this)); -941this.writestate.flushRequested = false; -942this.writestate.compacting.set(0); -943 -944if (this.writestate.writesEnabled) { -945 LOG.debug("Cleaning up temporary data for " + this.getRegionInfo().getEncodedName()); -946 // Remove temporary data left over from old regions -947 status.setStatus("Cleaning up temporary data from old regions"); -948 fs.cleanupTempDir(); -949} -950 -951if (this.writestate.writesEnabled) { -952 status.setStatus("Cleaning up detritus from prior splits"); -953 // Get rid of any splits or merges that were lost in-progress. Clean out -954 // these directories here on open. We may be opening a region that was -955 // being split but we crashed in the middle of it all. -956 LOG.debug("Cleaning up detritus for " + this.getRegionInfo().getEncodedName()); -957 fs.cleanupAnySplitDetritus(); -958 fs.cleanupMergesDir(); -959} +880 @VisibleForTesting +881 long initialize(final CancelableProgressable reporter) throws IOException { +882 +883//Refuse to open the region if there is no column family in the table +884if (htableDescriptor.getColumnFamilyCount() == 0) { +885 throw new DoNotRetryIOException("Table " + htableDescriptor.getTableName().getNameAsString()+ +886 " should have at least one column family.");
[20/51] [partial] hbase-site git commit: Published site at 63f2d3cbdc8151f5f61f33e0a078c51b9ac076a5.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ae6a80c/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html deleted file mode 100644 index a8b7714..000 --- a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStates.RegionStateNode.html +++ /dev/null @@ -1,539 +0,0 @@ -http://www.w3.org/TR/html4/loose.dtd;> - - - - - -Uses of Class org.apache.hadoop.hbase.master.assignment.RegionStates.RegionStateNode (Apache HBase 3.0.0-SNAPSHOT API) - - - - - - - -JavaScript is disabled on your browser. - - - - - -Skip navigation links - - - - -Overview -Package -Class -Use -Tree -Deprecated -Index -Help - - - - -Prev -Next - - -Frames -NoFrames - - -AllClasses - - - - - - - - - - -Uses of Classorg.apache.hadoop.hbase.master.assignment.RegionStates.RegionStateNode - - - - - -Packages that use RegionStates.RegionStateNode - -Package -Description - - - -org.apache.hadoop.hbase.master.assignment - - - - - - - - - - -Uses of RegionStates.RegionStateNode in org.apache.hadoop.hbase.master.assignment - -Fields in org.apache.hadoop.hbase.master.assignment declared as RegionStates.RegionStateNode - -Modifier and Type -Field and Description - - - -private RegionStates.RegionStateNode -RegionStates.RegionFailedOpen.regionNode - - - - -Fields in org.apache.hadoop.hbase.master.assignment with type parameters of type RegionStates.RegionStateNode - -Modifier and Type -Field and Description - - - -private https://docs.oracle.com/javase/8/docs/api/java/util/ArrayList.html?is-external=true; title="class or interface in java.util">ArrayListRegionStates.RegionStateNode -AssignmentManager.pendingAssignQueue - - -private https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentSkipListMap.html?is-external=true; title="class or interface in java.util.concurrent">ConcurrentSkipListMapRegionInfo,RegionStates.RegionStateNode -RegionStates.regionInTransition - - -private https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentSkipListMap.html?is-external=true; title="class or interface in java.util.concurrent">ConcurrentSkipListMapRegionInfo,RegionStates.RegionStateNode -RegionStates.regionOffline -Regions marked as offline on a read of hbase:meta. - - - -private https://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">SetRegionStates.RegionStateNode -RegionStates.ServerStateNode.regions - - -private https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentSkipListMap.html?is-external=true; title="class or interface in java.util.concurrent">ConcurrentSkipListMapbyte[],RegionStates.RegionStateNode -RegionStates.regionsMap -RegionName -- i.e. - - - - - -Methods in org.apache.hadoop.hbase.master.assignment that return RegionStates.RegionStateNode - -Modifier and Type -Method and Description - - - -protected RegionStates.RegionStateNode -RegionStates.createRegionStateNode(RegionInforegionInfo) - - -protected RegionStates.RegionStateNode -RegionStates.getOrCreateRegionStateNode(RegionInforegionInfo) - - -RegionStates.RegionStateNode -RegionTransitionProcedure.getRegionState(MasterProcedureEnvenv) - - -RegionStates.RegionStateNode -RegionStates.RegionFailedOpen.getRegionStateNode() - - -protected RegionStates.RegionStateNode -RegionStates.getRegionStateNode(RegionInforegionInfo) - - -(package private) RegionStates.RegionStateNode -RegionStates.getRegionStateNodeFromName(byte[]regionName) - - - - -Methods in org.apache.hadoop.hbase.master.assignment that return types with arguments of type RegionStates.RegionStateNode - -Modifier and Type -Method and Description - - - -https://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">SetRegionStates.RegionStateNode -RegionStates.ServerStateNode.getRegions() - - -https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionStates.RegionStateNode -RegionStates.getRegionsInTransition() - - -https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in
[20/51] [partial] hbase-site git commit: Published site at 092efb42749bf7fc6ad338c96aae8e7b9d3a2c74.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f3d62514/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html index 6086d40..3cfacfc 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/Procedure.html @@ -903,103 +903,108 @@ 895 return; 896} 897 -898LOG.debug("{} held the lock before restarting, call acquireLock to restore it.", this); -899LockState state = acquireLock(env); -900assert state == LockState.LOCK_ACQUIRED; -901 } +898if (isFinished()) { +899 LOG.debug("{} is already finished, skip acquiring lock.", this); +900 return; +901} 902 -903 /** -904 * Internal method called by the ProcedureExecutor that starts the user-level code acquireLock(). -905 */ -906 final LockState doAcquireLock(TEnvironment env, ProcedureStore store) { -907if (waitInitialized(env)) { -908 return LockState.LOCK_EVENT_WAIT; -909} -910if (lockedWhenLoading) { -911 // reset it so we will not consider it anymore -912 lockedWhenLoading = false; -913 locked = true; -914 // Here we return without persist the locked state, as lockedWhenLoading is true means -915 // that the locked field of the procedure stored in procedure store is true, so we do not need -916 // to store it again. -917 return LockState.LOCK_ACQUIRED; -918} -919LockState state = acquireLock(env); -920if (state == LockState.LOCK_ACQUIRED) { -921 locked = true; -922 // persist that we have held the lock. This must be done before we actually execute the -923 // procedure, otherwise when restarting, we may consider the procedure does not have a lock, -924 // but it may have already done some changes as we have already executed it, and if another -925 // procedure gets the lock, then the semantic will be broken if the holdLock is true, as we do -926 // not expect that another procedure can be executed in the middle. -927 store.update(this); -928} -929return state; -930 } -931 -932 /** -933 * Internal method called by the ProcedureExecutor that starts the user-level code releaseLock(). -934 */ -935 final void doReleaseLock(TEnvironment env, ProcedureStore store) { -936locked = false; -937// persist that we have released the lock. This must be done before we actually release the -938// lock. Another procedure may take this lock immediately after we release the lock, and if we -939// crash before persist the information that we have already released the lock, then when -940// restarting there will be two procedures which both have the lock and cause problems. -941if (getState() != ProcedureState.ROLLEDBACK) { -942 // If the state is ROLLEDBACK, it means that we have already deleted the procedure from -943 // procedure store, so do not need to log the release operation any more. -944 store.update(this); -945} -946releaseLock(env); -947 } -948 -949 @Override -950 public int compareTo(final ProcedureTEnvironment other) { -951return Long.compare(getProcId(), other.getProcId()); +903LOG.debug("{} held the lock before restarting, call acquireLock to restore it.", this); +904LockState state = acquireLock(env); +905assert state == LockState.LOCK_ACQUIRED; +906 } +907 +908 /** +909 * Internal method called by the ProcedureExecutor that starts the user-level code acquireLock(). +910 */ +911 final LockState doAcquireLock(TEnvironment env, ProcedureStore store) { +912if (waitInitialized(env)) { +913 return LockState.LOCK_EVENT_WAIT; +914} +915if (lockedWhenLoading) { +916 // reset it so we will not consider it anymore +917 lockedWhenLoading = false; +918 locked = true; +919 // Here we return without persist the locked state, as lockedWhenLoading is true means +920 // that the locked field of the procedure stored in procedure store is true, so we do not need +921 // to store it again. +922 return LockState.LOCK_ACQUIRED; +923} +924LockState state = acquireLock(env); +925if (state == LockState.LOCK_ACQUIRED) { +926 locked = true; +927 // persist that we have held the lock. This must be done before we actually execute the +928 // procedure, otherwise when restarting, we may consider the procedure does not have a lock, +929 // but it may have already done some changes as we have already executed it, and if another +930 // procedure gets the lock, then the semantic will be broken if the holdLock is true, as we do +931 // not expect that another procedure can be executed in the middle. +932
[20/51] [partial] hbase-site git commit: Published site at 613d831429960348dc42c3bdb6ea5d31be15c81c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntry.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntry.html b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntry.html index bd3c59e..21e240a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntry.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntry.html @@ -33,62 +33,62 @@ 025import java.io.FileNotFoundException; 026import java.io.FileOutputStream; 027import java.io.IOException; -028import java.io.ObjectInputStream; -029import java.io.ObjectOutputStream; -030import java.io.Serializable; -031import java.nio.ByteBuffer; -032import java.util.ArrayList; -033import java.util.Comparator; -034import java.util.HashSet; -035import java.util.Iterator; -036import java.util.List; -037import java.util.Map; -038import java.util.NavigableSet; -039import java.util.PriorityQueue; -040import java.util.Set; -041import java.util.concurrent.ArrayBlockingQueue; -042import java.util.concurrent.BlockingQueue; -043import java.util.concurrent.ConcurrentHashMap; -044import java.util.concurrent.ConcurrentMap; -045import java.util.concurrent.ConcurrentSkipListSet; -046import java.util.concurrent.Executors; -047import java.util.concurrent.ScheduledExecutorService; -048import java.util.concurrent.TimeUnit; -049import java.util.concurrent.atomic.AtomicInteger; -050import java.util.concurrent.atomic.AtomicLong; -051import java.util.concurrent.atomic.LongAdder; -052import java.util.concurrent.locks.Lock; -053import java.util.concurrent.locks.ReentrantLock; -054import java.util.concurrent.locks.ReentrantReadWriteLock; -055import org.apache.hadoop.conf.Configuration; -056import org.apache.hadoop.hbase.HBaseConfiguration; -057import org.apache.hadoop.hbase.io.HeapSize; -058import org.apache.hadoop.hbase.io.hfile.BlockCache; -059import org.apache.hadoop.hbase.io.hfile.BlockCacheKey; -060import org.apache.hadoop.hbase.io.hfile.BlockCacheUtil; -061import org.apache.hadoop.hbase.io.hfile.BlockPriority; -062import org.apache.hadoop.hbase.io.hfile.BlockType; -063import org.apache.hadoop.hbase.io.hfile.CacheStats; -064import org.apache.hadoop.hbase.io.hfile.Cacheable; -065import org.apache.hadoop.hbase.io.hfile.Cacheable.MemoryType; -066import org.apache.hadoop.hbase.io.hfile.CacheableDeserializer; -067import org.apache.hadoop.hbase.io.hfile.CacheableDeserializerIdManager; -068import org.apache.hadoop.hbase.io.hfile.CachedBlock; -069import org.apache.hadoop.hbase.io.hfile.HFileBlock; -070import org.apache.hadoop.hbase.nio.ByteBuff; -071import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -072import org.apache.hadoop.hbase.util.HasThread; -073import org.apache.hadoop.hbase.util.IdReadWriteLock; -074import org.apache.hadoop.hbase.util.IdReadWriteLock.ReferenceType; -075import org.apache.hadoop.hbase.util.UnsafeAvailChecker; -076import org.apache.hadoop.util.StringUtils; -077import org.apache.yetus.audience.InterfaceAudience; -078import org.slf4j.Logger; -079import org.slf4j.LoggerFactory; -080 -081import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; -082import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; -083import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; +028import java.io.Serializable; +029import java.nio.ByteBuffer; +030import java.util.ArrayList; +031import java.util.Comparator; +032import java.util.HashSet; +033import java.util.Iterator; +034import java.util.List; +035import java.util.Map; +036import java.util.NavigableSet; +037import java.util.PriorityQueue; +038import java.util.Set; +039import java.util.concurrent.ArrayBlockingQueue; +040import java.util.concurrent.BlockingQueue; +041import java.util.concurrent.ConcurrentHashMap; +042import java.util.concurrent.ConcurrentMap; +043import java.util.concurrent.ConcurrentSkipListSet; +044import java.util.concurrent.Executors; +045import java.util.concurrent.ScheduledExecutorService; +046import java.util.concurrent.TimeUnit; +047import java.util.concurrent.atomic.AtomicInteger; +048import java.util.concurrent.atomic.AtomicLong; +049import java.util.concurrent.atomic.LongAdder; +050import java.util.concurrent.locks.Lock; +051import java.util.concurrent.locks.ReentrantLock; +052import java.util.concurrent.locks.ReentrantReadWriteLock; +053import org.apache.hadoop.conf.Configuration; +054import org.apache.hadoop.hbase.HBaseConfiguration; +055import org.apache.hadoop.hbase.io.HeapSize; +056import org.apache.hadoop.hbase.io.hfile.BlockCache; +057import org.apache.hadoop.hbase.io.hfile.BlockCacheKey; +058import org.apache.hadoop.hbase.io.hfile.BlockCacheUtil;
[20/51] [partial] hbase-site git commit: Published site at ba5d1c1f28301adc99019d9d6c4a04fac98ae511.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/804782f0/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html b/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html index 1b3ae53..6c4852a 100644 --- a/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html +++ b/devapidocs/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html @@ -186,7 +186,7 @@ extends CLEANER_INTERVAL_CONF_KEY -private https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">Long,ProcedureExecutor.CompletedProcedureRetainer +private https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">Long,ProcedureExecutor.CompletedProcedureRetainerTEnvironment completed @@ -253,7 +253,7 @@ extends CompletedProcedureCleaner(org.apache.hadoop.conf.Configurationconf, ProcedureStorestore, - https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">Long,ProcedureExecutor.CompletedProcedureRetainercompletedMap, + https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">Long,ProcedureExecutor.CompletedProcedureRetainerTEnvironmentcompletedMap, https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">MapNonceKey,https://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true; title="class or interface in java.lang">LongnonceKeysToProcIdsMap) @@ -288,7 +288,7 @@ extends Procedure -acquireLock, addStackIndex, afterReplay, beforeReplay, compareTo, completionCleanup, doAcquireLock, doExecute, doReleaseLock, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, isYieldAfterExecutionStep, releaseLock, removeStackIndex, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, shouldWaitClientAck, toString, toS tringClass, toStringClassDetails, toStringDetails, toStringSimpleSB, toStringState, tryRunnable, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, wasExecuted +acquireLock, addStackIndex, afterReplay, beforeReplay, compareTo, completionCleanup, doAcquireLock, doExecute, doReleaseLock, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, holdLock, incChildrenLatch, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, isYieldAfterExecutionStep, lockedWhenLoading, releaseLock, removeStackIndex, restoreLock, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParent ProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailu re, shouldWaitClientAck, toString, toStringClass, toStringClassDetails, toStringDetails, toStringSimpleSB, toStringState, tryRunnable, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, waitInitialized, wasExecuted @@ -430,7 +430,7 @@ extends
[20/51] [partial] hbase-site git commit: Published site at b4759ce6e72f50ccd9d410bd5917dc5a515414f1.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/df8fd1d3/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html index 233dba3..91b9055 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MoveCostFunction.html @@ -540,1205 +540,1204 @@ 532 sm.getRegionMetrics().forEach((byte[] regionName, RegionMetrics rm) - { 533DequeBalancerRegionLoad rLoads = oldLoads.get(Bytes.toString(regionName)); 534if (rLoads == null) { -535 // There was nothing there -536 rLoads = new ArrayDeque(); -537} else if (rLoads.size() = numRegionLoadsToRemember) { -538 rLoads.remove(); -539} -540rLoads.add(new BalancerRegionLoad(rm)); -541 loads.put(Bytes.toString(regionName), rLoads); -542 }); -543}); -544 -545for(CostFromRegionLoadFunction cost : regionLoadFunctions) { -546 cost.setLoads(loads); -547} -548 } -549 -550 protected void initCosts(Cluster cluster) { -551for (CostFunction c:costFunctions) { -552 c.init(cluster); -553} -554 } -555 -556 protected void updateCostsWithAction(Cluster cluster, Action action) { -557for (CostFunction c : costFunctions) { -558 c.postAction(action); -559} -560 } -561 -562 /** -563 * Get the names of the cost functions -564 */ -565 public String[] getCostFunctionNames() { -566if (costFunctions == null) return null; -567String[] ret = new String[costFunctions.length]; -568for (int i = 0; i costFunctions.length; i++) { -569 CostFunction c = costFunctions[i]; -570 ret[i] = c.getClass().getSimpleName(); -571} -572 -573return ret; -574 } -575 -576 /** -577 * This is the main cost function. It will compute a cost associated with a proposed cluster -578 * state. All different costs will be combined with their multipliers to produce a double cost. -579 * -580 * @param cluster The state of the cluster -581 * @param previousCost the previous cost. This is used as an early out. -582 * @return a double of a cost associated with the proposed cluster state. This cost is an -583 * aggregate of all individual cost functions. -584 */ -585 protected double computeCost(Cluster cluster, double previousCost) { -586double total = 0; -587 -588for (int i = 0; i costFunctions.length; i++) { -589 CostFunction c = costFunctions[i]; -590 this.tempFunctionCosts[i] = 0.0; -591 -592 if (c.getMultiplier() = 0) { -593continue; -594 } -595 -596 Float multiplier = c.getMultiplier(); -597 Double cost = c.cost(); -598 -599 this.tempFunctionCosts[i] = multiplier*cost; -600 total += this.tempFunctionCosts[i]; -601 -602 if (total previousCost) { -603break; -604 } -605} -606 -607return total; -608 } -609 -610 /** Generates a candidate action to be applied to the cluster for cost function search */ -611 abstract static class CandidateGenerator { -612abstract Cluster.Action generate(Cluster cluster); -613 -614/** -615 * From a list of regions pick a random one. Null can be returned which -616 * {@link StochasticLoadBalancer#balanceCluster(Map)} recognize as signal to try a region move -617 * rather than swap. -618 * -619 * @param clusterThe state of the cluster -620 * @param server index of the server -621 * @param chanceOfNoSwap Chance that this will decide to try a move rather -622 * than a swap. -623 * @return a random {@link RegionInfo} or null if an asymmetrical move is -624 * suggested. -625 */ -626protected int pickRandomRegion(Cluster cluster, int server, double chanceOfNoSwap) { -627 // Check to see if this is just a move. -628 if (cluster.regionsPerServer[server].length == 0 || RANDOM.nextFloat() chanceOfNoSwap) { -629// signal a move only. -630return -1; -631 } -632 int rand = RANDOM.nextInt(cluster.regionsPerServer[server].length); -633 return cluster.regionsPerServer[server][rand]; -634 -635} -636protected int pickRandomServer(Cluster cluster) { -637 if (cluster.numServers 1) { -638return -1; -639 } -640 -641 return RANDOM.nextInt(cluster.numServers); -642} -643 -644protected int pickRandomRack(Cluster cluster) { -645 if (cluster.numRacks 1) { -646return -1; -647 } -648 -649 return
[20/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0c6f447e/apidocs/org/apache/hadoop/hbase/RetryImmediatelyException.html -- diff --git a/apidocs/org/apache/hadoop/hbase/RetryImmediatelyException.html b/apidocs/org/apache/hadoop/hbase/RetryImmediatelyException.html index 08c03f4..36052a7 100644 --- a/apidocs/org/apache/hadoop/hbase/RetryImmediatelyException.html +++ b/apidocs/org/apache/hadoop/hbase/RetryImmediatelyException.html @@ -1,6 +1,6 @@ http://www.w3.org/TR/html4/loose.dtd;> - + @@ -20,38 +20,38 @@ //--> -æ¨çæµè§å¨å·²ç¦ç¨ JavaScriptã +JavaScript is disabled on your browser. -è·³è¿å¯¼èªé¾æ¥ +Skip navigation links - -æ¦è§ -ç¨åºå -ç±» -ä½¿ç¨ -æ -å·²è¿æ¶ -ç´¢å¼ -å¸®å© + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help -ä¸ä¸ä¸ªç±» -ä¸ä¸ä¸ªç±» +PrevClass +NextClass -æ¡æ¶ -æ æ¡æ¶ +Frames +NoFrames -ææç±» +AllClassesorg.apache.hadoop.hbase-ç±» RetryImmediatelyException
+Class RetryImmediatelyException
-
- https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true" title="java.langä¸çç±»ææ¥å£">java.lang.Object
+- https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true" title="class or interface in java.lang">java.lang.Object
-
- https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true" title="java.langä¸çç±»ææ¥å£">java.lang.Throwable
+- https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true" title="class or interface in java.lang">java.lang.Throwable
-
- https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true" title="java.langä¸çç±»ææ¥å£">java.lang.Exception
+- https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true" title="class or interface in java.lang">java.lang.Exception
-
- https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="java.ioä¸çç±»ææ¥å£">java.io.IOException
+- https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true" title="class or interface in java.io">java.io.IOException
- org.apache.hadoop.hbase.RetryImmediatelyException
@@ -117,21 +117,21 @@
-
- ææå·²å®ç°çæ¥å£:
-
[20/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5427a45e/apidocs/org/apache/hadoop/hbase/RegionMetrics.html -- diff --git a/apidocs/org/apache/hadoop/hbase/RegionMetrics.html b/apidocs/org/apache/hadoop/hbase/RegionMetrics.html index c745b96..e4dc75c 100644 --- a/apidocs/org/apache/hadoop/hbase/RegionMetrics.html +++ b/apidocs/org/apache/hadoop/hbase/RegionMetrics.html @@ -1,6 +1,6 @@ http://www.w3.org/TR/html4/loose.dtd;> - + @@ -19,45 +19,45 @@ } //--> var methods = {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":6,"i7":6,"i8":6,"i9":18,"i10":6,"i11":6,"i12":18,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6,"i20":6,"i21":6}; -var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],16:["t5","Default Methods"]}; +var tabs = {65535:["t0","æææ¹æ³"],2:["t2","å®ä¾æ¹æ³"],4:["t3","æ½è±¡æ¹æ³"],16:["t5","é»è®¤æ¹æ³"]}; var altColor = "altColor"; var rowColor = "rowColor"; var tableTab = "tableTab"; var activeTableTab = "activeTableTab"; -JavaScript is disabled on your browser. +æ¨çæµè§å¨å·²ç¦ç¨ JavaScriptã -Skip navigation links +è·³è¿å¯¼èªé¾æ¥ - -Overview -Package -Class -Use -Tree -Deprecated -Index -Help + +æ¦è§ +ç¨åºå +ç±» +ä½¿ç¨ +æ +å·²è¿æ¶ +ç´¢å¼ +å¸®å© -PrevClass -NextClass +ä¸ä¸ä¸ªç±» +ä¸ä¸ä¸ªç±» -Frames -NoFrames +æ¡æ¶ +æ æ¡æ¶ -AllClasses +ææç±»org.apache.hadoop.hbase-Interface RegionMetrics
+æ¥å£ RegionMetrics
-
- All Known Implementing Classes:
-- RegionLoad
+- ææå·²ç¥å®ç°ç±»:
+- RegionLoad
@@ -119,15 +119,15 @@ public interface -Method Summary
-