[21/51] [partial] hbase-site git commit: Published site at 620d70d6186fb800299bcc62ad7179fccfd1be41.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa3fb87f/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html index 68302bf..a5a8905 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html @@ -2197,1768 +2197,1775 @@ 2189 warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null); 2190} 2191 -2192for (ColumnFamilyDescriptor hcd : htd.getColumnFamilies()) { -2193 if (hcd.getTimeToLive() = 0) { -2194String message = "TTL for column family " + hcd.getNameAsString() + " must be positive."; -2195 warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null); -2196 } -2197 -2198 // check blockSize -2199 if (hcd.getBlocksize() 1024 || hcd.getBlocksize() 16 * 1024 * 1024) { -2200String message = "Block size for column family " + hcd.getNameAsString() -2201+ " must be between 1K and 16MB."; +2192// check that we have minimum 1 region replicas +2193int regionReplicas = htd.getRegionReplication(); +2194if (regionReplicas 1) { +2195 String message = "Table region replication should be at least one."; +2196 warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null); +2197} +2198 +2199for (ColumnFamilyDescriptor hcd : htd.getColumnFamilies()) { +2200 if (hcd.getTimeToLive() = 0) { +2201String message = "TTL for column family " + hcd.getNameAsString() + " must be positive."; 2202 warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null); 2203 } 2204 -2205 // check versions -2206 if (hcd.getMinVersions() 0) { -2207String message = "Min versions for column family " + hcd.getNameAsString() -2208 + " must be positive."; +2205 // check blockSize +2206 if (hcd.getBlocksize() 1024 || hcd.getBlocksize() 16 * 1024 * 1024) { +2207String message = "Block size for column family " + hcd.getNameAsString() +2208+ " must be between 1K and 16MB."; 2209 warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null); 2210 } -2211 // max versions already being checked -2212 -2213 // HBASE-13776 Setting illegal versions for ColumnFamilyDescriptor -2214 // does not throw IllegalArgumentException -2215 // check minVersions = maxVerions -2216 if (hcd.getMinVersions() hcd.getMaxVersions()) { -2217String message = "Min versions for column family " + hcd.getNameAsString() -2218+ " must be less than the Max versions."; -2219 warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null); -2220 } -2221 - // check replication scope -2223 checkReplicationScope(hcd); -2224 // check bloom filter type -2225 checkBloomFilterType(hcd); -2226 -2227 // check data replication factor, it can be 0(default value) when user has not explicitly -2228 // set the value, in this case we use default replication factor set in the file system. -2229 if (hcd.getDFSReplication() 0) { -2230String message = "HFile Replication for column family " + hcd.getNameAsString() -2231+ " must be greater than zero."; -2232 warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null); -2233 } -2234 -2235 // TODO: should we check coprocessors and encryption ? -2236} -2237 } -2238 -2239 private void checkReplicationScope(ColumnFamilyDescriptor hcd) throws IOException{ -2240// check replication scope -2241WALProtos.ScopeType scop = WALProtos.ScopeType.valueOf(hcd.getScope()); -2242if (scop == null) { -2243 String message = "Replication scope for column family " -2244 + hcd.getNameAsString() + " is " + hcd.getScope() + " which is invalid."; +2211 +2212 // check versions +2213 if (hcd.getMinVersions() 0) { +2214String message = "Min versions for column family " + hcd.getNameAsString() +2215 + " must be positive."; +2216 warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null); +2217 } +2218 // max versions already being checked +2219 +2220 // HBASE-13776 Setting illegal versions for ColumnFamilyDescriptor +2221 // does not throw IllegalArgumentException + // check minVersions = maxVerions +2223 if (hcd.getMinVersions() hcd.getMaxVersions()) { +2224String message = "Min versions for column family " + hcd.getNameAsString() +2225+ " must be less than the Max versions."; +2226 warnOrThrowExceptionForFailure(logWarn, CONF_KEY, message, null); +2227 } +2228 +2229 // check replication scope +2230
[21/51] [partial] hbase-site git commit: Published site at 281d6429e55149cc4c05430dcc1d1dc136d8b245.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/901d593a/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.html b/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.html index bd7effb..9a61b35 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.html +++ b/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i 109":10,"i110":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i 109":10,"i110":10,"i111":10,"i112":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -119,7 +119,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public class MasterRpcServices +public class MasterRpcServices extends RSRpcServices implements org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterService.BlockingInterface, org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStatusService.BlockingInterface, org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockService.BlockingInterface, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.HbckService.BlockingInterface Implements the master RPC services. @@ -559,142 +559,147 @@ implements org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.Master +org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse +isRpcThrottleEnabled(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequestrequest) + + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse isSnapshotDone(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequestrequest) Checks if the specified snapshot is done. - + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse isSplitOrMergeEnabled(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequestrequest) - + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse listDecommissionedRegionServers(org.apache.hbase.thirdparty.com.google.protobuf.RpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequestrequest) - + org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse
[21/51] [partial] hbase-site git commit: Published site at 466fa920fee572fe20db3b77ebf539dc304d5f31.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2bf59208/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder.html index 2e150bc..0b315b8 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder.html @@ -25,22 +25,22 @@ 017 */ 018package org.apache.hadoop.hbase.client; 019 -020import static org.apache.hbase.thirdparty.com.google.common.base.Preconditions.checkArgument; -021import static org.apache.hbase.thirdparty.com.google.common.base.Preconditions.checkNotNull; -022import static org.apache.hadoop.hbase.client.ConnectionUtils.retries2Attempts; +020import static org.apache.hadoop.hbase.client.ConnectionUtils.retries2Attempts; +021import static org.apache.hbase.thirdparty.com.google.common.base.Preconditions.checkArgument; +022import static org.apache.hbase.thirdparty.com.google.common.base.Preconditions.checkNotNull; 023 -024import org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer; -025 -026import java.util.List; -027import java.util.concurrent.CompletableFuture; -028import java.util.concurrent.TimeUnit; -029 -030import org.apache.hadoop.hbase.HRegionLocation; -031import org.apache.hadoop.hbase.ServerName; -032import org.apache.hadoop.hbase.TableName; -033import org.apache.yetus.audience.InterfaceAudience; -034import org.apache.hadoop.hbase.client.metrics.ScanMetrics; -035import org.apache.hadoop.hbase.ipc.HBaseRpcController; +024import java.util.List; +025import java.util.concurrent.CompletableFuture; +026import java.util.concurrent.TimeUnit; +027import org.apache.hadoop.hbase.HRegionLocation; +028import org.apache.hadoop.hbase.ServerName; +029import org.apache.hadoop.hbase.TableName; +030import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +031import org.apache.hadoop.hbase.ipc.HBaseRpcController; +032import org.apache.yetus.audience.InterfaceAudience; +033 +034import org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer; +035 036import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ClientService; 037import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ScanResponse; 038 @@ -83,432 +83,441 @@ 075 076private RegionLocateType locateType = RegionLocateType.CURRENT; 077 -078public SingleRequestCallerBuilderT table(TableName tableName) { -079 this.tableName = tableName; -080 return this; -081} -082 -083public SingleRequestCallerBuilderT row(byte[] row) { -084 this.row = row; -085 return this; -086} -087 -088public SingleRequestCallerBuilderT action( -089 AsyncSingleRequestRpcRetryingCaller.CallableT callable) { -090 this.callable = callable; -091 return this; -092} -093 -094public SingleRequestCallerBuilderT operationTimeout(long operationTimeout, TimeUnit unit) { -095 this.operationTimeoutNs = unit.toNanos(operationTimeout); -096 return this; -097} -098 -099public SingleRequestCallerBuilderT rpcTimeout(long rpcTimeout, TimeUnit unit) { -100 this.rpcTimeoutNs = unit.toNanos(rpcTimeout); -101 return this; -102} -103 -104public SingleRequestCallerBuilderT locateType(RegionLocateType locateType) { -105 this.locateType = locateType; -106 return this; -107} -108 -109public SingleRequestCallerBuilderT pause(long pause, TimeUnit unit) { -110 this.pauseNs = unit.toNanos(pause); -111 return this; -112} -113 -114public SingleRequestCallerBuilderT maxAttempts(int maxAttempts) { -115 this.maxAttempts = maxAttempts; -116 return this; -117} -118 -119public SingleRequestCallerBuilderT startLogErrorsCnt(int startLogErrorsCnt) { -120 this.startLogErrorsCnt = startLogErrorsCnt; -121 return this; -122} -123 -124public AsyncSingleRequestRpcRetryingCallerT build() { -125 return new AsyncSingleRequestRpcRetryingCaller(retryTimer, conn, -126 checkNotNull(tableName, "tableName is null"), checkNotNull(row, "row is null"), -127 checkNotNull(locateType, "locateType is null"), checkNotNull(callable, "action is null"), -128 pauseNs, maxAttempts, operationTimeoutNs, rpcTimeoutNs, startLogErrorsCnt); +078private int replicaId = RegionReplicaUtil.DEFAULT_REPLICA_ID; +079 +080public SingleRequestCallerBuilderT table(TableName tableName) { +081 this.tableName = tableName; +082 return this; +083} +084 +085public SingleRequestCallerBuilderT row(byte[] row)
[21/51] [partial] hbase-site git commit: Published site at e4b6b4afb933a961f543537875f87a2dc62d3757.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/849d84a8/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftHBaseServiceHandler.IOErrorWithCause.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftHBaseServiceHandler.IOErrorWithCause.html b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftHBaseServiceHandler.IOErrorWithCause.html new file mode 100644 index 000..5b5b199 --- /dev/null +++ b/devapidocs/src-html/org/apache/hadoop/hbase/thrift/ThriftHBaseServiceHandler.IOErrorWithCause.html @@ -0,0 +1,1419 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + +Source code + + + + +001/** +002 * +003 * Licensed to the Apache Software Foundation (ASF) under one +004 * or more contributor license agreements. See the NOTICE file +005 * distributed with this work for additional information +006 * regarding copyright ownership. The ASF licenses this file +007 * to you under the Apache License, Version 2.0 (the +008 * "License"); you may not use this file except in compliance +009 * with the License. You may obtain a copy of the License at +010 * +011 * http://www.apache.org/licenses/LICENSE-2.0 +012 * +013 * Unless required by applicable law or agreed to in writing, software +014 * distributed under the License is distributed on an "AS IS" BASIS, +015 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +016 * See the License for the specific language governing permissions and +017 * limitations under the License. +018 */ +019 +020package org.apache.hadoop.hbase.thrift; +021 +022import static org.apache.hadoop.hbase.thrift.Constants.COALESCE_INC_KEY; +023import static org.apache.hadoop.hbase.util.Bytes.getBytes; +024 +025import java.io.IOException; +026import java.nio.ByteBuffer; +027import java.util.ArrayList; +028import java.util.Collections; +029import java.util.HashMap; +030import java.util.List; +031import java.util.Map; +032import java.util.TreeMap; +033 +034import org.apache.hadoop.conf.Configuration; +035import org.apache.hadoop.hbase.Cell; +036import org.apache.hadoop.hbase.CellBuilder; +037import org.apache.hadoop.hbase.CellBuilderFactory; +038import org.apache.hadoop.hbase.CellBuilderType; +039import org.apache.hadoop.hbase.CellUtil; +040import org.apache.hadoop.hbase.HColumnDescriptor; +041import org.apache.hadoop.hbase.HConstants; +042import org.apache.hadoop.hbase.HRegionLocation; +043import org.apache.hadoop.hbase.HTableDescriptor; +044import org.apache.hadoop.hbase.KeyValue; +045import org.apache.hadoop.hbase.MetaTableAccessor; +046import org.apache.hadoop.hbase.ServerName; +047import org.apache.hadoop.hbase.TableName; +048import org.apache.hadoop.hbase.TableNotFoundException; +049import org.apache.hadoop.hbase.client.Append; +050import org.apache.hadoop.hbase.client.Delete; +051import org.apache.hadoop.hbase.client.Durability; +052import org.apache.hadoop.hbase.client.Get; +053import org.apache.hadoop.hbase.client.Increment; +054import org.apache.hadoop.hbase.client.OperationWithAttributes; +055import org.apache.hadoop.hbase.client.Put; +056import org.apache.hadoop.hbase.client.RegionInfo; +057import org.apache.hadoop.hbase.client.RegionLocator; +058import org.apache.hadoop.hbase.client.Result; +059import org.apache.hadoop.hbase.client.ResultScanner; +060import org.apache.hadoop.hbase.client.Scan; +061import org.apache.hadoop.hbase.client.Table; +062import org.apache.hadoop.hbase.filter.Filter; +063import org.apache.hadoop.hbase.filter.ParseFilter; +064import org.apache.hadoop.hbase.filter.PrefixFilter; +065import org.apache.hadoop.hbase.filter.WhileMatchFilter; +066import org.apache.hadoop.hbase.security.UserProvider; +067import org.apache.hadoop.hbase.thrift.generated.AlreadyExists; +068import org.apache.hadoop.hbase.thrift.generated.BatchMutation; +069import org.apache.hadoop.hbase.thrift.generated.ColumnDescriptor; +070import org.apache.hadoop.hbase.thrift.generated.Hbase; +071import org.apache.hadoop.hbase.thrift.generated.IOError; +072import org.apache.hadoop.hbase.thrift.generated.IllegalArgument; +073import org.apache.hadoop.hbase.thrift.generated.Mutation; +074import org.apache.hadoop.hbase.thrift.generated.TAppend; +075import org.apache.hadoop.hbase.thrift.generated.TCell; +076import org.apache.hadoop.hbase.thrift.generated.TIncrement; +077import org.apache.hadoop.hbase.thrift.generated.TRegionInfo; +078import org.apache.hadoop.hbase.thrift.generated.TRowResult; +079import org.apache.hadoop.hbase.thrift.generated.TScan; +080import org.apache.hadoop.hbase.util.Bytes; +081import org.apache.thrift.TException; +082import org.apache.yetus.audience.InterfaceAudience; +083import org.slf4j.Logger; +084import org.slf4j.LoggerFactory; +085 +086import org.apache.hbase.thirdparty.com.google.common.base.Throwables; +087 +088/** +089 * The HBaseServiceHandler is a glue object that connects Thrift
[21/51] [partial] hbase-site git commit: Published site at 3ab895979b643a2980bcdb7fee2078f14b614210.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/coprocessor/example/RowCountEndpoint.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/RowCountEndpoint.html b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/RowCountEndpoint.html index cb536db..261e004 100644 --- a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/RowCountEndpoint.html +++ b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/RowCountEndpoint.html @@ -479,6 +479,6 @@ implements Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ScanModifyingObserver.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ScanModifyingObserver.html b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ScanModifyingObserver.html index 17f90e2..47cd2d5 100644 --- a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ScanModifyingObserver.html +++ b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ScanModifyingObserver.html @@ -470,6 +470,6 @@ implements Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ValueRewritingObserver.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ValueRewritingObserver.html b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ValueRewritingObserver.html index da0a020..b7f2a4f 100644 --- a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ValueRewritingObserver.html +++ b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ValueRewritingObserver.html @@ -511,6 +511,6 @@ implements Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/coprocessor/example/WriteHeavyIncrementObserver.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/WriteHeavyIncrementObserver.html b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/WriteHeavyIncrementObserver.html index c7c92c6..edd3ed4 100644 --- a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/WriteHeavyIncrementObserver.html +++ b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/WriteHeavyIncrementObserver.html @@ -788,6 +788,6 @@ implements Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.ZKDataHolder.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.ZKDataHolder.html b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.ZKDataHolder.html index b7d8b2f..c136a98 100644 --- a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.ZKDataHolder.html +++ b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.ZKDataHolder.html @@ -411,6 +411,6 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html -Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. +Copyright 20072019 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ef0dd56d/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.html b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.html index a8d8e27..cd7ee0e 100644 --- a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.html +++
[21/51] [partial] hbase-site git commit: Published site at 7820ba1dbdba58b1002cdfde08eb21aa7a0bb6da.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/09ea0d5f/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftUtilities.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftUtilities.html b/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftUtilities.html index 62da309..c7ea445 100644 --- a/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftUtilities.html +++ b/devapidocs/org/apache/hadoop/hbase/thrift2/ThriftUtilities.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9,"i7":9,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":9,"i19":9}; +var methods = {"i0":9,"i1":9,"i2":9,"i3":9,"i4":9,"i5":9,"i6":9,"i7":9,"i8":9,"i9":9,"i10":9,"i11":9,"i12":9,"i13":9,"i14":9,"i15":9,"i16":9,"i17":9,"i18":9,"i19":9,"i20":9,"i21":9,"i22":9,"i23":9,"i24":9,"i25":9,"i26":9,"i27":9,"i28":9,"i29":9,"i30":9,"i31":9,"i32":9,"i33":9,"i34":9,"i35":9,"i36":9,"i37":9,"i38":9,"i39":9,"i40":9,"i41":9,"i42":9,"i43":9}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -75,13 +75,13 @@ var activeTableTab = "activeTableTab"; Summary: Nested| -Field| +Field| Constr| Method Detail: -Field| +Field| Constr| Method @@ -110,7 +110,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private -public final class ThriftUtilities +public final class ThriftUtilities extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object @@ -118,6 +118,33 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html + + + + + +Field Summary + +Fields + +Modifier and Type +Field and Description + + +private static Cell[] +EMPTY_CELL_ARRAY + + +private static Result +EMPTY_RESULT + + +private static Result +EMPTY_RESULT_STALE + + + + @@ -161,95 +188,191 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html appendFromThrift(org.apache.hadoop.hbase.thrift2.generated.TAppendappend) +static org.apache.hadoop.hbase.thrift2.generated.TBloomFilterType +bloomFilterFromHBase(BloomTypein) + + +static BloomType +bloomFilterFromThrift(org.apache.hadoop.hbase.thrift2.generated.TBloomFilterTypein) + + +static org.apache.hadoop.hbase.thrift2.generated.TColumnFamilyDescriptor +columnFamilyDescriptorFromHBase(ColumnFamilyDescriptorin) + + +static ColumnFamilyDescriptor +columnFamilyDescriptorFromThrift(org.apache.hadoop.hbase.thrift2.generated.TColumnFamilyDescriptorin) + + static CompareOperator compareOpFromThrift(org.apache.hadoop.hbase.thrift2.generated.TCompareOptCompareOp) - + +static org.apache.hadoop.hbase.thrift2.generated.TCompressionAlgorithm +compressionAlgorithmFromHBase(Compression.Algorithmin) + + +static Compression.Algorithm +compressionAlgorithmFromThrift(org.apache.hadoop.hbase.thrift2.generated.TCompressionAlgorithmin) + + private static Consistency consistencyFromThrift(org.apache.hadoop.hbase.thrift2.generated.TConsistencytConsistency) - + +static org.apache.hadoop.hbase.thrift2.generated.TDataBlockEncoding +dataBlockEncodingFromHBase(DataBlockEncodingin) + + +static DataBlockEncoding +dataBlockEncodingFromThrift(org.apache.hadoop.hbase.thrift2.generated.TDataBlockEncodingin) + + static org.apache.hadoop.hbase.thrift2.generated.TDelete deleteFromHBase(Deletein) - + static Delete deleteFromThrift(org.apache.hadoop.hbase.thrift2.generated.TDeletein) Creates a Delete (HBase) from a TDelete (Thrift). - + static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListDelete deletesFromThrift(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.hbase.thrift2.generated.TDeletein) Converts multiple TDeletes (Thrift) into a list of Deletes (HBase). - + +static org.apache.hadoop.hbase.thrift2.generated.TDeleteType +deleteTypeFromHBase(Cell.Typetype) + + +private static org.apache.hadoop.hbase.thrift2.generated.TDurability +durabilityFromHBase(Durabilitydurability) + + private static Durability durabilityFromThrift(org.apache.hadoop.hbase.thrift2.generated.TDurabilitytDurability) - + static Get getFromThrift(org.apache.hadoop.hbase.thrift2.generated.TGetin) Creates a Get (HBase) from a TGet (Thrift). - + static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListGet getsFromThrift(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.hbase.thrift2.generated.TGetin) Converts multiple TGets (Thrift) into a list of Gets (HBase). - + static
[21/51] [partial] hbase-site git commit: Published site at c448604ceb987d113913f0583452b2abce04db0d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServerServices.PostOpenDeployContext.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServerServices.PostOpenDeployContext.html b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServerServices.PostOpenDeployContext.html index bfe700b..bd404ad 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServerServices.PostOpenDeployContext.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServerServices.PostOpenDeployContext.html @@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab"; -public static class RegionServerServices.PostOpenDeployContext +public static class RegionServerServices.PostOpenDeployContext extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object Context for postOpenDeployTasks(). @@ -211,7 +211,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html region -private finalHRegion region +private finalHRegion region @@ -220,7 +220,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html masterSystemTime -private finallong masterSystemTime +private finallong masterSystemTime @@ -238,7 +238,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html PostOpenDeployContext @InterfaceAudience.Private -publicPostOpenDeployContext(HRegionregion, +publicPostOpenDeployContext(HRegionregion, longmasterSystemTime) @@ -256,7 +256,7 @@ public getRegion -publicHRegiongetRegion() +publicHRegiongetRegion() @@ -265,7 +265,7 @@ public getMasterSystemTime -publiclonggetMasterSystemTime() +publiclonggetMasterSystemTime() http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4f8b8424/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServerServices.RegionStateTransitionContext.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServerServices.RegionStateTransitionContext.html b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServerServices.RegionStateTransitionContext.html index c642a6b..84b5eca 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServerServices.RegionStateTransitionContext.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/RegionServerServices.RegionStateTransitionContext.html @@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab"; -public static class RegionServerServices.RegionStateTransitionContext +public static class RegionServerServices.RegionStateTransitionContext extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object @@ -228,7 +228,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html code -private finalorg.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition.TransitionCode code +private finalorg.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition.TransitionCode code @@ -237,7 +237,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html openSeqNum -private finallong openSeqNum +private finallong openSeqNum @@ -246,7 +246,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html masterSystemTime -private finallong masterSystemTime +private finallong masterSystemTime @@ -255,7 +255,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html hris -private finalRegionInfo[] hris +private finalRegionInfo[] hris @@ -273,7 +273,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html RegionStateTransitionContext @InterfaceAudience.Private -publicRegionStateTransitionContext(org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition.TransitionCodecode, +publicRegionStateTransitionContext(org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition.TransitionCodecode, longopenSeqNum, longmasterSystemTime, RegionInfo...hris) @@ -293,7 +293,7 @@ public getCode -publicorg.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition.TransitionCodegetCode() +publicorg.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition.TransitionCodegetCode() @@ -302,7 +302,7 @@ public getOpenSeqNum -publiclonggetOpenSeqNum()
[21/51] [partial] hbase-site git commit: Published site at 8bf966c8e936dec4d83bcbe85c5aab543f14a0df.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/27555316/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.RSGroupStartupWorker.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.RSGroupStartupWorker.html b/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.RSGroupStartupWorker.html index 809f66f..9b60dd0 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.RSGroupStartupWorker.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.RSGroupStartupWorker.html @@ -765,146 +765,145 @@ 757found.set(true); 758try { 759 boolean rootMetaFound = -760 masterServices.getMetaTableLocator().verifyMetaRegionLocation( -761 conn, masterServices.getZooKeeper(), 1); -762 if (rootMetaFound) { -763MetaTableAccessor.Visitor visitor = new DefaultVisitorBase() { -764 @Override -765 public boolean visitInternal(Result row) throws IOException { -766RegionInfo info = MetaTableAccessor.getRegionInfo(row); -767if (info != null) { -768 Cell serverCell = -769 row.getColumnLatestCell(HConstants.CATALOG_FAMILY, -770 HConstants.SERVER_QUALIFIER); -771 if (RSGROUP_TABLE_NAME.equals(info.getTable()) serverCell != null) { -772ServerName sn = -773 ServerName.parseVersionedServerName(CellUtil.cloneValue(serverCell)); -774if (sn == null) { -775 found.set(false); -776} else if (tsm.isTableState(RSGROUP_TABLE_NAME, TableState.State.ENABLED)) { -777 try { -778 ClientProtos.ClientService.BlockingInterface rs = conn.getClient(sn); -779 ClientProtos.GetRequest request = -780 RequestConverter.buildGetRequest(info.getRegionName(), -781new Get(ROW_KEY)); -782rs.get(null, request); -783 assignedRegions.add(info); -784 } catch(Exception ex) { -785LOG.debug("Caught exception while verifying group region", ex); -786 } -787} -788 foundRegions.add(info); -789 } -790} -791return true; -792 } -793}; -794 MetaTableAccessor.fullScanRegions(conn, visitor); -795// if no regions in meta then we have to create the table -796if (foundRegions.size() 1 rootMetaFound !createSent) { -797 createRSGroupTable(); -798 createSent = true; -799} -800LOG.info("RSGroup table=" + RSGROUP_TABLE_NAME + " isOnline=" + found.get() -801+ ", regionCount=" + foundRegions.size() + ", assignCount=" -802+ assignedRegions.size() + ", rootMetaFound=" + rootMetaFound); -803found.set(found.get() assignedRegions.size() == foundRegions.size() -804 foundRegions.size() 0); -805 } else { -806LOG.info("Waiting for catalog tables to come online"); -807found.set(false); -808 } -809 if (found.get()) { -810LOG.debug("With group table online, refreshing cached information."); -811 RSGroupInfoManagerImpl.this.refresh(true); -812online = true; -813//flush any inconsistencies between ZK and HTable -814 RSGroupInfoManagerImpl.this.flushConfig(); -815 } -816} catch (RuntimeException e) { -817 throw e; -818} catch(Exception e) { -819 found.set(false); -820 LOG.warn("Failed to perform check", e); -821} -822try { -823 Thread.sleep(100); -824} catch (InterruptedException e) { -825 LOG.info("Sleep interrupted", e); -826} -827 } -828 return found.get(); -829} -830 -831private void createRSGroupTable() throws IOException { -832 Long procId = masterServices.createSystemTable(RSGROUP_TABLE_DESC); -833 // wait for region to be online -834 int tries = 600; -835 while (!(masterServices.getMasterProcedureExecutor().isFinished(procId)) -836 masterServices.getMasterProcedureExecutor().isRunning() -837 tries 0) { -838try { -839 Thread.sleep(100); -840} catch (InterruptedException e) { -841 throw new IOException("Wait interrupted ", e); -842} -843tries--; -844 } -845 if(tries = 0) { -846throw new
[21/51] [partial] hbase-site git commit: Published site at 1acbd36c903b048141866b143507bfce124a5c5f.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5299e667/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html index 883e8b2..b6da844 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html @@ -1172,6 +1172,13 @@ InitMetaProcedure.acquireLock(MasterProcedureEnvenv) +protected static void +AbstractStateMachineNamespaceProcedure.addOrUpdateNamespace(MasterProcedureEnvenv, +NamespaceDescriptorns) +Insert/update the row into the ns family of meta table. + + + private void RestoreSnapshotProcedure.addRegionsToInMemoryStates(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInforegionInfos, MasterProcedureEnvenv, @@ -1179,13 +1186,13 @@ Add regions to in-memory states - + private void CloneSnapshotProcedure.addRegionsToMeta(MasterProcedureEnvenv) Add regions to hbase:meta table. - + private static void CreateTableProcedure.addRegionsToMeta(MasterProcedureEnvenv, TableDescriptortableDescriptor, @@ -1193,60 +1200,60 @@ Add the specified set of regions to the hbase:meta table. - + private static void ModifyTableProcedure.addRegionsToMeta(MasterProcedureEnvenv, TableDescriptortableDescriptor, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInforegionInfos) - + protected static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo CreateTableProcedure.addTableToMeta(MasterProcedureEnvenv, TableDescriptortableDescriptor, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInforegions) - + private void ServerCrashProcedure.assignRegions(MasterProcedureEnvenv, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInforegions) Assign the regions on the crashed RS to other Rses. - + private static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionRequest RSProcedureDispatcher.buildOpenRegionRequest(MasterProcedureEnvenv, ServerNameserverName, https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRSProcedureDispatcher.RegionOpenOperationoperations) - + org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo RSProcedureDispatcher.RegionOpenOperation.buildRegionOpenInfoRequest(MasterProcedureEnvenv) - + private boolean ReopenTableRegionsProcedure.canSchedule(MasterProcedureEnvenv, HRegionLocationloc) - + protected static void AbstractStateMachineTableProcedure.checkOnline(MasterProcedureEnvenv, RegionInfori) Check region is online. - + protected void AbstractStateMachineRegionProcedure.checkTableModifiable(MasterProcedureEnvenv) Check whether a table is modifiable - exists and either offline or online with config set - + protected void AbstractStateMachineTableProcedure.checkTableModifiable(MasterProcedureEnvenv) Check whether a table is modifiable - exists and either offline or online with config set - + private static void DeleteTableProcedure.cleanAnyRemainingRows(MasterProcedureEnvenv, TableNametableName) @@ -1254,26 +1261,26 @@ info:regioninfo column was empty because of some write error. - + protected void TruncateTableProcedure.completionCleanup(MasterProcedureEnvenv) - + protected void ModifyTableProcedure.completionCleanup(MasterProcedureEnvenv) - + protected void InitMetaProcedure.completionCleanup(MasterProcedureEnvenv) - + protected static void -CreateNamespaceProcedure.createDirectory(MasterProcedureEnvenv, +AbstractStateMachineNamespaceProcedure.createDirectory(MasterProcedureEnvenv, NamespaceDescriptornsDescriptor) Create the namespace directory - + private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionInfo CloneSnapshotProcedure.createFilesystemLayout(MasterProcedureEnvenv, TableDescriptortableDescriptor, @@ -1281,20 +1288,20 @@ Create regions in file system. - + protected static
[21/51] [partial] hbase-site git commit: Published site at 130057f13774f6b213cdb06952c805a29d59396e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/68eae623/devapidocs/src-html/org/apache/hadoop/hbase/security/access/AuthManager.PermissionCache.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/security/access/AuthManager.PermissionCache.html b/devapidocs/src-html/org/apache/hadoop/hbase/security/access/AuthManager.PermissionCache.html new file mode 100644 index 000..4d5cbc9 --- /dev/null +++ b/devapidocs/src-html/org/apache/hadoop/hbase/security/access/AuthManager.PermissionCache.html @@ -0,0 +1,680 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + +Source code + + + + +001/** +002 * Licensed to the Apache Software Foundation (ASF) under one +003 * or more contributor license agreements. See the NOTICE file +004 * distributed with this work for additional information +005 * regarding copyright ownership. The ASF licenses this file +006 * to you under the Apache License, Version 2.0 (the +007 * "License"); you may not use this file except in compliance +008 * with the License. You may obtain a copy of the License at +009 * +010 * http://www.apache.org/licenses/LICENSE-2.0 +011 * +012 * Unless required by applicable law or agreed to in writing, software +013 * distributed under the License is distributed on an "AS IS" BASIS, +014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +015 * See the License for the specific language governing permissions and +016 * limitations under the License. +017 */ +018 +019package org.apache.hadoop.hbase.security.access; +020 +021import java.io.Closeable; +022import java.io.IOException; +023import java.util.HashMap; +024import java.util.HashSet; +025import java.util.List; +026import java.util.Map; +027import java.util.Set; +028import java.util.concurrent.ConcurrentHashMap; +029import java.util.concurrent.atomic.AtomicLong; +030 +031import org.apache.hadoop.conf.Configuration; +032import org.apache.hadoop.hbase.AuthUtil; +033import org.apache.hadoop.hbase.Cell; +034import org.apache.hadoop.hbase.TableName; +035import org.apache.hadoop.hbase.exceptions.DeserializationException; +036import org.apache.hadoop.hbase.log.HBaseMarkers; +037import org.apache.hadoop.hbase.security.Superusers; +038import org.apache.hadoop.hbase.security.User; +039import org.apache.hadoop.hbase.security.UserProvider; +040import org.apache.hadoop.hbase.util.Bytes; +041import org.apache.hadoop.hbase.zookeeper.ZKWatcher; +042import org.apache.yetus.audience.InterfaceAudience; +043import org.apache.zookeeper.KeeperException; +044import org.slf4j.Logger; +045import org.slf4j.LoggerFactory; +046 +047import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; +048import org.apache.hbase.thirdparty.com.google.common.collect.ListMultimap; +049import org.apache.hbase.thirdparty.com.google.common.collect.Lists; +050 +051/** +052 * Performs authorization checks for a given user's assigned permissions. +053 * p +054 * There're following scopes: bGlobal/b, bNamespace/b, bTable/b, bFamily/b, +055 * bQualifier/b, bCell/b. +056 * Generally speaking, higher scopes can overrides lower scopes, +057 * except for Cell permission can be granted even a user has not permission on specified table, +058 * which means the user can get/scan only those granted cells parts. +059 * /p +060 * e.g, if user A has global permission R(ead), he can +061 * read table T without checking table scope permission, so authorization checks alway starts from +062 * Global scope. +063 * p +064 * For each scope, not only user but also groups he belongs to will be checked. +065 * /p +066 */ +067@InterfaceAudience.Private +068public final class AuthManager implements Closeable { +069 +070 /** +071 * Cache of permissions, it is thread safe. +072 * @param T T extends Permission +073 */ +074 private static class PermissionCacheT extends Permission { +075private final Object mutex = new Object(); +076private MapString, SetT cache = new HashMap(); +077 +078void put(String name, T perm) { +079 synchronized (mutex) { +080SetT perms = cache.getOrDefault(name, new HashSet()); +081perms.add(perm); +082cache.put(name, perms); +083 } +084} +085 +086SetT get(String name) { +087 synchronized (mutex) { +088return cache.get(name); +089 } +090} +091 +092void clear() { +093 synchronized (mutex) { +094for (Map.EntryString, SetT entry : cache.entrySet()) { +095 entry.getValue().clear(); +096} +097cache.clear(); +098 } +099} +100 } +101 PermissionCacheNamespacePermission NS_NO_PERMISSION = new PermissionCache(); +102 PermissionCacheTablePermission TBL_NO_PERMISSION = new PermissionCache(); +103 +104 /** +105 * Cache for global permission. +106 * Since every user/group can only have one global permission, no need to user
[21/51] [partial] hbase-site git commit: Published site at d5e4faacc354c1bc4d93efa71ca97ee3a056123e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b5e107c3/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html index b2a9771..bf81ebb 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionsCleaner.html @@ -46,3768 +46,3806 @@ 038import java.util.Objects; 039import java.util.Set; 040import java.util.SortedMap; -041import java.util.TreeMap; -042import java.util.TreeSet; -043import java.util.concurrent.ConcurrentHashMap; -044import java.util.concurrent.ConcurrentMap; -045import java.util.concurrent.ConcurrentSkipListMap; -046import java.util.concurrent.atomic.AtomicBoolean; -047import java.util.concurrent.locks.ReentrantReadWriteLock; -048import java.util.function.Function; -049import javax.management.MalformedObjectNameException; -050import javax.servlet.http.HttpServlet; -051import org.apache.commons.lang3.RandomUtils; -052import org.apache.commons.lang3.StringUtils; -053import org.apache.commons.lang3.SystemUtils; -054import org.apache.hadoop.conf.Configuration; -055import org.apache.hadoop.fs.FileSystem; -056import org.apache.hadoop.fs.Path; -057import org.apache.hadoop.hbase.Abortable; -058import org.apache.hadoop.hbase.CacheEvictionStats; -059import org.apache.hadoop.hbase.ChoreService; -060import org.apache.hadoop.hbase.ClockOutOfSyncException; -061import org.apache.hadoop.hbase.CoordinatedStateManager; -062import org.apache.hadoop.hbase.DoNotRetryIOException; -063import org.apache.hadoop.hbase.HBaseConfiguration; -064import org.apache.hadoop.hbase.HBaseInterfaceAudience; -065import org.apache.hadoop.hbase.HConstants; -066import org.apache.hadoop.hbase.HealthCheckChore; -067import org.apache.hadoop.hbase.MetaTableAccessor; -068import org.apache.hadoop.hbase.NotServingRegionException; -069import org.apache.hadoop.hbase.PleaseHoldException; -070import org.apache.hadoop.hbase.ScheduledChore; -071import org.apache.hadoop.hbase.ServerName; -072import org.apache.hadoop.hbase.Stoppable; -073import org.apache.hadoop.hbase.TableDescriptors; -074import org.apache.hadoop.hbase.TableName; -075import org.apache.hadoop.hbase.YouAreDeadException; -076import org.apache.hadoop.hbase.ZNodeClearer; -077import org.apache.hadoop.hbase.client.ClusterConnection; -078import org.apache.hadoop.hbase.client.Connection; -079import org.apache.hadoop.hbase.client.ConnectionUtils; -080import org.apache.hadoop.hbase.client.RegionInfo; -081import org.apache.hadoop.hbase.client.RegionInfoBuilder; -082import org.apache.hadoop.hbase.client.RpcRetryingCallerFactory; -083import org.apache.hadoop.hbase.client.TableDescriptorBuilder; -084import org.apache.hadoop.hbase.client.locking.EntityLock; -085import org.apache.hadoop.hbase.client.locking.LockServiceClient; -086import org.apache.hadoop.hbase.conf.ConfigurationManager; -087import org.apache.hadoop.hbase.conf.ConfigurationObserver; -088import org.apache.hadoop.hbase.coordination.SplitLogWorkerCoordination; -089import org.apache.hadoop.hbase.coordination.ZkCoordinatedStateManager; -090import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; -091import org.apache.hadoop.hbase.exceptions.RegionMovedException; -092import org.apache.hadoop.hbase.exceptions.RegionOpeningException; -093import org.apache.hadoop.hbase.exceptions.UnknownProtocolException; -094import org.apache.hadoop.hbase.executor.ExecutorService; -095import org.apache.hadoop.hbase.executor.ExecutorType; -096import org.apache.hadoop.hbase.fs.HFileSystem; -097import org.apache.hadoop.hbase.http.InfoServer; -098import org.apache.hadoop.hbase.io.hfile.BlockCache; -099import org.apache.hadoop.hbase.io.hfile.CacheConfig; -100import org.apache.hadoop.hbase.io.hfile.HFile; -101import org.apache.hadoop.hbase.io.util.MemorySizeUtil; -102import org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils; -103import org.apache.hadoop.hbase.ipc.NettyRpcClientConfigHelper; -104import org.apache.hadoop.hbase.ipc.RpcClient; -105import org.apache.hadoop.hbase.ipc.RpcClientFactory; -106import org.apache.hadoop.hbase.ipc.RpcControllerFactory; -107import org.apache.hadoop.hbase.ipc.RpcServer; -108import org.apache.hadoop.hbase.ipc.RpcServerInterface; -109import org.apache.hadoop.hbase.ipc.ServerNotRunningYetException; -110import org.apache.hadoop.hbase.ipc.ServerRpcController; -111import org.apache.hadoop.hbase.log.HBaseMarkers; -112import org.apache.hadoop.hbase.master.HMaster; -113import org.apache.hadoop.hbase.master.LoadBalancer; -114import org.apache.hadoop.hbase.master.RegionState.State; -115import
[21/51] [partial] hbase-site git commit: Published site at 3fe8649b2c9ba1271c25e8f476548907e4c7a90d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8f09a71d/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html index c7d99b2..9d1542c 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.RegionReplicaCandidateGenerator.html @@ -382,1357 +382,1365 @@ 374for (int i = 0; i this.curFunctionCosts.length; i++) { 375 curFunctionCosts[i] = tempFunctionCosts[i]; 376} -377LOG.info("start StochasticLoadBalancer.balancer, initCost=" + currentCost + ", functionCost=" -378+ functionCost()); +377double initCost = currentCost; +378double newCost = currentCost; 379 -380double initCost = currentCost; -381double newCost = currentCost; -382 -383long computedMaxSteps; -384if (runMaxSteps) { -385 computedMaxSteps = Math.max(this.maxSteps, -386 ((long)cluster.numRegions * (long)this.stepsPerRegion * (long)cluster.numServers)); -387} else { -388 computedMaxSteps = Math.min(this.maxSteps, -389 ((long)cluster.numRegions * (long)this.stepsPerRegion * (long)cluster.numServers)); -390} -391// Perform a stochastic walk to see if we can get a good fit. -392long step; -393 -394for (step = 0; step computedMaxSteps; step++) { -395 Cluster.Action action = nextAction(cluster); -396 -397 if (action.type == Type.NULL) { -398continue; -399 } -400 -401 cluster.doAction(action); -402 updateCostsWithAction(cluster, action); -403 -404 newCost = computeCost(cluster, currentCost); -405 -406 // Should this be kept? -407 if (newCost currentCost) { -408currentCost = newCost; -409 -410// save for JMX -411curOverallCost = currentCost; -412for (int i = 0; i this.curFunctionCosts.length; i++) { -413 curFunctionCosts[i] = tempFunctionCosts[i]; -414} -415 } else { -416// Put things back the way they were before. -417// TODO: undo by remembering old values -418Action undoAction = action.undoAction(); -419cluster.doAction(undoAction); -420updateCostsWithAction(cluster, undoAction); -421 } -422 -423 if (EnvironmentEdgeManager.currentTime() - startTime -424 maxRunningTime) { -425break; -426 } -427} -428long endTime = EnvironmentEdgeManager.currentTime(); -429 -430 metricsBalancer.balanceCluster(endTime - startTime); -431 -432// update costs metrics -433updateStochasticCosts(tableName, curOverallCost, curFunctionCosts); -434if (initCost currentCost) { -435 plans = createRegionPlans(cluster); -436 LOG.info("Finished computing new load balance plan. Computation took {}" + -437" to try {} different iterations. Found a solution that moves " + -438"{} regions; Going from a computed cost of {}" + -439" to a new cost of {}", java.time.Duration.ofMillis(endTime - startTime), -440step, plans.size(), initCost, currentCost); -441 return plans; -442} -443LOG.info("Could not find a better load balance plan. Tried {} different configurations in " + -444 "{}, and did not find anything with a computed cost less than {}", step, -445 java.time.Duration.ofMillis(endTime - startTime), initCost); -446return null; -447 } -448 -449 /** -450 * update costs to JMX -451 */ -452 private void updateStochasticCosts(TableName tableName, Double overall, Double[] subCosts) { -453if (tableName == null) return; -454 -455// check if the metricsBalancer is MetricsStochasticBalancer before casting -456if (metricsBalancer instanceof MetricsStochasticBalancer) { -457 MetricsStochasticBalancer balancer = (MetricsStochasticBalancer) metricsBalancer; -458 // overall cost -459 balancer.updateStochasticCost(tableName.getNameAsString(), -460"Overall", "Overall cost", overall); -461 -462 // each cost function -463 for (int i = 0; i costFunctions.length; i++) { -464CostFunction costFunction = costFunctions[i]; -465String costFunctionName = costFunction.getClass().getSimpleName(); -466Double costPercent = (overall == 0) ? 0 : (subCosts[i] / overall); -467// TODO: cost function may need a specific description -468 balancer.updateStochasticCost(tableName.getNameAsString(), costFunctionName, -469 "The percent of " + costFunctionName,
[21/51] [partial] hbase-site git commit: Published site at 7adf590106826b9e4432cfeee06acdc0ccff8c6e.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/425db230/testdevapidocs/org/apache/hadoop/hbase/HBaseTestingUtility.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/HBaseTestingUtility.html b/testdevapidocs/org/apache/hadoop/hbase/HBaseTestingUtility.html index 9f9d0e1..b2fc784 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/HBaseTestingUtility.html +++ b/testdevapidocs/org/apache/hadoop/hbase/HBaseTestingUtility.html @@ -124,7 +124,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Public -public class HBaseTestingUtility +public class HBaseTestingUtility extends HBaseZKTestingUtility Facility for testing HBase. Replacement for old HBaseTestCase and HBaseClusterTestCase functionality. @@ -140,7 +140,8 @@ extends To preserve test data directories, pass the system property "hbase.testing.preserve.testdir" - setting it to true. + setting it to true. + Trigger pre commit. @@ -2018,7 +2019,7 @@ extends TEST_DIRECTORY_KEY https://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true; title="class or interface in java.lang">@Deprecated -private static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String TEST_DIRECTORY_KEY +private static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String TEST_DIRECTORY_KEY Deprecated.can be used only with mini dfs System property key to get test directory value. Name is as it is because mini dfs has hard-codings to put test data here. It should NOT be used directly in HBase, as it's a property @@ -2035,7 +2036,7 @@ private static finalhttps://docs.oracle.com/javase/8/docs/api/jav REGIONS_PER_SERVER_KEY -public static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String REGIONS_PER_SERVER_KEY +public static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String REGIONS_PER_SERVER_KEY See Also: Constant Field Values @@ -2048,7 +2049,7 @@ private static finalhttps://docs.oracle.com/javase/8/docs/api/jav DEFAULT_REGIONS_PER_SERVER -public static finalint DEFAULT_REGIONS_PER_SERVER +public static finalint DEFAULT_REGIONS_PER_SERVER The default number of regions per regionserver when creating a pre-split table. @@ -2063,7 +2064,7 @@ private static finalhttps://docs.oracle.com/javase/8/docs/api/jav PRESPLIT_TEST_TABLE_KEY -public static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String PRESPLIT_TEST_TABLE_KEY +public static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String PRESPLIT_TEST_TABLE_KEY See Also: Constant Field Values @@ -2076,7 +2077,7 @@ private static finalhttps://docs.oracle.com/javase/8/docs/api/jav PRESPLIT_TEST_TABLE -public static finalboolean PRESPLIT_TEST_TABLE +public static finalboolean PRESPLIT_TEST_TABLE See Also: Constant Field Values @@ -2089,7 +2090,7 @@ private static finalhttps://docs.oracle.com/javase/8/docs/api/jav dfsCluster -privateorg.apache.hadoop.hdfs.MiniDFSCluster dfsCluster +privateorg.apache.hadoop.hdfs.MiniDFSCluster dfsCluster @@ -2098,7 +2099,7 @@ private static finalhttps://docs.oracle.com/javase/8/docs/api/jav hbaseCluster -private volatileHBaseCluster hbaseCluster +private volatileHBaseCluster hbaseCluster @@ -2107,7 +2108,7 @@ private static finalhttps://docs.oracle.com/javase/8/docs/api/jav mrCluster -privateorg.apache.hadoop.mapred.MiniMRCluster mrCluster +privateorg.apache.hadoop.mapred.MiniMRCluster mrCluster @@ -2116,7 +2117,7 @@ private static finalhttps://docs.oracle.com/javase/8/docs/api/jav miniClusterRunning -private volatileboolean miniClusterRunning +private volatileboolean miniClusterRunning If there is a mini cluster running for this testing utility instance. @@ -2126,7 +2127,7 @@ private static finalhttps://docs.oracle.com/javase/8/docs/api/jav hadoopLogDir -privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String hadoopLogDir +privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String hadoopLogDir @@ -2135,7 +2136,7 @@ private static finalhttps://docs.oracle.com/javase/8/docs/api/jav dataTestDirOnTestFS -privateorg.apache.hadoop.fs.Path dataTestDirOnTestFS +privateorg.apache.hadoop.fs.Path dataTestDirOnTestFS Directory on test filesystem where we put the data for this instance of HBaseTestingUtility @@ -2146,7 +2147,7 @@
[21/51] [partial] hbase-site git commit: Published site at 5fbb227deb365fe812d433fe39b85ac4b0ddee20.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c9ebe686/devapidocs/org/apache/hadoop/hbase/master/procedure/TruncateTableProcedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/procedure/TruncateTableProcedure.html b/devapidocs/org/apache/hadoop/hbase/master/procedure/TruncateTableProcedure.html index 093bc8e..7744c8c 100644 --- a/devapidocs/org/apache/hadoop/hbase/master/procedure/TruncateTableProcedure.html +++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/TruncateTableProcedure.html @@ -377,7 +377,7 @@ extends Procedure -addStackIndex, afterReplay, beforeReplay, compareTo, doExecute, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProcIdHashCode, get ProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, incChildrenLatch, isBypass, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, removeStackIndex, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, shouldWaitClientAck, skipPersistence, toString, toStringClass, toStringDetails, toStringSimpleSB, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, wasExecuted +addStackIndex, afterReplay, beforeReplay, bypass, compareTo, doExecute, doRollback, elapsedTime, getChildrenLatch, getException, getLastUpdate, getNonceKey, getOwner, getParentProcId, getProcedureMetrics, getProcId, getProc IdHashCode, getProcName, getResult, getRootProcedureId, getRootProcId, getStackIndexes, getState, getSubmittedTime, getTimeout, getTimeoutTimestamp, hasChildren, hasException, hasLock, hasOwner, hasParent, hasTimeout, haveSameParent, incChildrenLatch, isBypass, isFailed, isFinished, isInitializing, isRunnable, isSuccess, isWaiting, removeStackIndex, setAbortFailure, setChildrenLatch, setFailure, setFailure, setLastUpdate, setNonceKey, setOwner, setOwner, setParentProcId, setProcId, setResult, setRootProcId, setStackIndexes, setState, setSubmittedTime, setTimeout, setTimeoutFailure, shouldWaitClientAck, skipPersistence, toString, toStringClass, toStringDetails, toStringSimpleSB, updateMetricsOnFinish, updateMetricsOnSubmit, updateTimestamp, wasExecuted
[21/51] [partial] hbase-site git commit: Published site at 821e4d7de2d576189f4288d1c2acf9e9a9471f5c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/323b17d9/testdevapidocs/org/apache/hadoop/hbase/client/TestRestoreSnapshotFromClientAfterSplittingRegions.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/client/TestRestoreSnapshotFromClientAfterSplittingRegions.html b/testdevapidocs/org/apache/hadoop/hbase/client/TestRestoreSnapshotFromClientAfterSplittingRegions.html new file mode 100644 index 000..cb4690a --- /dev/null +++ b/testdevapidocs/org/apache/hadoop/hbase/client/TestRestoreSnapshotFromClientAfterSplittingRegions.html @@ -0,0 +1,369 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +TestRestoreSnapshotFromClientAfterSplittingRegions (Apache HBase 3.0.0-SNAPSHOT Test API) + + + + + +var methods = {"i0":10,"i1":9}; +var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Field| +Constr| +Method + + +Detail: +Field| +Constr| +Method + + + + + + + + +org.apache.hadoop.hbase.client +Class TestRestoreSnapshotFromClientAfterSplittingRegions + + + +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +org.apache.hadoop.hbase.client.RestoreSnapshotFromClientTestBase + + +org.apache.hadoop.hbase.client.RestoreSnapshotFromClientAfterSplittingRegionsTestBase + + +org.apache.hadoop.hbase.client.TestRestoreSnapshotFromClientAfterSplittingRegions + + + + + + + + + + + + +public class TestRestoreSnapshotFromClientAfterSplittingRegions +extends RestoreSnapshotFromClientAfterSplittingRegionsTestBase + + + + + + + + + + + +Field Summary + +Fields + +Modifier and Type +Field and Description + + +static HBaseClassTestRule +CLASS_RULE + + +int +numReplicas + + + + + + +Fields inherited from classorg.apache.hadoop.hbase.client.RestoreSnapshotFromClientTestBase +admin, emptySnapshot, FAMILY, name, snapshot0Rows, snapshot1Rows, snapshotName0, snapshotName1, snapshotName2, tableName, TEST_FAMILY2, TEST_UTIL + + + + + + + + +Constructor Summary + +Constructors + +Constructor and Description + + +TestRestoreSnapshotFromClientAfterSplittingRegions() + + + + + + + + + +Method Summary + +All MethodsStatic MethodsInstance MethodsConcrete Methods + +Modifier and Type +Method and Description + + +protected int +getNumReplicas() + + +static https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listhttps://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object[] +params() + + + + + + +Methods inherited from classorg.apache.hadoop.hbase.client.RestoreSnapshotFromClientAfterSplittingRegionsTestBase +testRestoreSnapshotAfterSplittingRegions + + + + + +Methods inherited from classorg.apache.hadoop.hbase.client.RestoreSnapshotFromClientTestBase +countRows, createTable, getValidMethodName, setup, setupCluster, setupConf, splitRe gion, tearDown, tearDownAfterClass, verifyRowCount + + + + + +Methods inherited from classjava.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--; title="class or interface in java.lang">clone, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-; title="class or interface in java.lang">equals, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--; title="class or interface in java.lang">finalize, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--; title="class or interface in java.lang">getClass, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--; title="class or interface in java.lang">hashCode, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--; title="class or interface in java.lang">notify,
[21/51] [partial] hbase-site git commit: Published site at fa5fa6ecdd071b72b58971058ff3ab9d28c3e709.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d1341859/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.ProcedureIterator.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.ProcedureIterator.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.ProcedureIterator.html index ff29160..e4dc134 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.ProcedureIterator.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/store/ProcedureStore.ProcedureIterator.html @@ -42,190 +42,208 @@ 034public interface ProcedureStore { 035 /** 036 * Store listener interface. -037 * The main process should register a listener and respond to the store events. -038 */ -039 public interface ProcedureStoreListener { -040/** -041 * triggered when the store sync is completed. -042 */ -043void postSync(); -044 -045/** -046 * triggered when the store is not able to write out data. -047 * the main process should abort. -048 */ -049void abortProcess(); -050 } -051 -052 /** -053 * An Iterator over a collection of Procedure -054 */ -055 public interface ProcedureIterator { -056/** -057 * Reset the Iterator by seeking to the beginning of the list. -058 */ -059void reset(); -060 -061/** -062 * Returns true if the iterator has more elements. -063 * (In other words, returns true if next() would return a Procedure -064 * rather than throwing an exception.) -065 * @return true if the iterator has more procedures -066 */ -067boolean hasNext(); -068 -069/** -070 * @return true if the iterator next element is a completed procedure. -071 */ -072boolean isNextFinished(); -073 +037 * p/ +038 * The main process should register a listener and respond to the store events. +039 */ +040 public interface ProcedureStoreListener { +041 +042/** +043 * triggered when the store sync is completed. +044 */ +045default void postSync() { +046} +047 +048/** +049 * triggered when the store is not able to write out data. the main process should abort. +050 */ +051default void abortProcess() { +052} +053 +054/** +055 * Suggest that the upper layer should update the state of some procedures. Ignore this call +056 * will not effect correctness but performance. +057 * p/ +058 * For a WAL based ProcedureStore implementation, if all the procedures stored in a WAL file +059 * have been deleted, or updated later in another WAL file, then we can delete the WAL file. If +060 * there are old procedures in a WAL file which are never deleted or updated, then we can not +061 * delete the WAL file and this will cause we hold lots of WAL file and slow down the master +062 * restarts. So here we introduce this method to tell the upper layer that please update the +063 * states of these procedures so that we can delete the old WAL file. +064 * @param procIds the id for the procedures +065 */ +066default void forceUpdate(long[] procIds) { +067} +068 } +069 +070 /** +071 * An Iterator over a collection of Procedure +072 */ +073 public interface ProcedureIterator { 074/** -075 * Skip the next procedure +075 * Reset the Iterator by seeking to the beginning of the list. 076 */ -077void skipNext(); +077void reset(); 078 079/** -080 * Returns the next procedure in the iteration. -081 * @throws IOException if there was an error fetching/deserializing the procedure -082 * @return the next procedure in the iteration. -083 */ -084@SuppressWarnings("rawtypes") -085Procedure next() throws IOException; -086 } -087 -088 /** -089 * Interface passed to the ProcedureStore.load() method to handle the store-load events. -090 */ -091 public interface ProcedureLoader { +080 * Returns true if the iterator has more elements. +081 * (In other words, returns true if next() would return a Procedure +082 * rather than throwing an exception.) +083 * @return true if the iterator has more procedures +084 */ +085boolean hasNext(); +086 +087/** +088 * @return true if the iterator next element is a completed procedure. +089 */ +090boolean isNextFinished(); +091 092/** -093 * Called by ProcedureStore.load() to notify about the maximum proc-id in the store. -094 * @param maxProcId the highest proc-id in the store -095 */ -096void setMaxProcId(long maxProcId); -097 -098/** -099 * Called by the ProcedureStore.load() every time a set of procedures are ready to be executed. -100 * The ProcedureIterator passed to the method, has the procedure sorted in replay-order. -101 * @param procIter iterator over
[21/51] [partial] hbase-site git commit: Published site at 6bc7089f9e0793efc9bdd46a84f5ccd9bc4579ad.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/419d0338/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html index 43c66a8..061ce80 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.Testing.html @@ -23,2136 +23,2142 @@ 015 * See the License for the specific language governing permissions and 016 * limitations under the License. 017 */ -018 -019package org.apache.hadoop.hbase.procedure2; -020 -021import java.io.IOException; -022import java.util.ArrayDeque; -023import java.util.ArrayList; -024import java.util.Arrays; -025import java.util.Collection; -026import java.util.Deque; -027import java.util.HashSet; -028import java.util.Iterator; -029import java.util.List; -030import java.util.Map; -031import java.util.Objects; -032import java.util.Set; -033import java.util.concurrent.ConcurrentHashMap; -034import java.util.concurrent.CopyOnWriteArrayList; -035import java.util.concurrent.TimeUnit; -036import java.util.concurrent.atomic.AtomicBoolean; -037import java.util.concurrent.atomic.AtomicInteger; -038import java.util.concurrent.atomic.AtomicLong; -039import java.util.stream.Collectors; -040import java.util.stream.Stream; -041 -042import org.apache.hadoop.conf.Configuration; -043import org.apache.hadoop.hbase.HConstants; -044import org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException; -045import org.apache.hadoop.hbase.log.HBaseMarkers; -046import org.apache.hadoop.hbase.procedure2.Procedure.LockState; -047import org.apache.hadoop.hbase.procedure2.store.ProcedureStore; -048import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator; -049import org.apache.hadoop.hbase.procedure2.util.StringUtils; -050import org.apache.hadoop.hbase.security.User; -051import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; -052import org.apache.hadoop.hbase.util.IdLock; -053import org.apache.hadoop.hbase.util.NonceKey; -054import org.apache.hadoop.hbase.util.Threads; -055import org.apache.yetus.audience.InterfaceAudience; -056import org.slf4j.Logger; -057import org.slf4j.LoggerFactory; -058 -059import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; -060import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; +018package org.apache.hadoop.hbase.procedure2; +019 +020import java.io.IOException; +021import java.util.ArrayDeque; +022import java.util.ArrayList; +023import java.util.Arrays; +024import java.util.Collection; +025import java.util.Deque; +026import java.util.HashSet; +027import java.util.Iterator; +028import java.util.List; +029import java.util.Map; +030import java.util.Objects; +031import java.util.Set; +032import java.util.concurrent.ConcurrentHashMap; +033import java.util.concurrent.CopyOnWriteArrayList; +034import java.util.concurrent.TimeUnit; +035import java.util.concurrent.atomic.AtomicBoolean; +036import java.util.concurrent.atomic.AtomicInteger; +037import java.util.concurrent.atomic.AtomicLong; +038import java.util.stream.Collectors; +039import java.util.stream.Stream; +040import org.apache.hadoop.conf.Configuration; +041import org.apache.hadoop.hbase.HConstants; +042import org.apache.hadoop.hbase.exceptions.IllegalArgumentIOException; +043import org.apache.hadoop.hbase.log.HBaseMarkers; +044import org.apache.hadoop.hbase.procedure2.Procedure.LockState; +045import org.apache.hadoop.hbase.procedure2.store.ProcedureStore; +046import org.apache.hadoop.hbase.procedure2.store.ProcedureStore.ProcedureIterator; +047import org.apache.hadoop.hbase.procedure2.util.StringUtils; +048import org.apache.hadoop.hbase.security.User; +049import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; +050import org.apache.hadoop.hbase.util.IdLock; +051import org.apache.hadoop.hbase.util.NonceKey; +052import org.apache.hadoop.hbase.util.Threads; +053import org.apache.yetus.audience.InterfaceAudience; +054import org.slf4j.Logger; +055import org.slf4j.LoggerFactory; +056 +057import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting; +058import org.apache.hbase.thirdparty.com.google.common.base.Preconditions; +059 +060import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState; 061 -062import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureState; -063 -064/** -065 * Thread Pool that executes the submitted procedures. -066 * The executor has a ProcedureStore associated. -067 * Each operation is logged and on restart the pending procedures are resumed. -068 * -069 * Unless the Procedure code throws an error
[21/51] [partial] hbase-site git commit: Published site at d7e08317d2f214e4cca7b67578aba0ed7a567d54.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/37cf49a6/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncAdmin.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncAdmin.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncAdmin.html index 77bb7f2..0019de7 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncAdmin.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncAdmin.html @@ -885,414 +885,417 @@ 877 MapString, String props); 878 879 /** -880 * abort a procedure -881 * @param procId ID of the procedure to abort -882 * @param mayInterruptIfRunning if the proc completed at least one step, should it be aborted? -883 * @return true if aborted, false if procedure already completed or does not exist. the value is -884 * wrapped by {@link CompletableFuture} -885 */ -886 CompletableFutureBoolean abortProcedure(long procId, boolean mayInterruptIfRunning); -887 -888 /** -889 * List procedures -890 * @return procedure list JSON wrapped by {@link CompletableFuture} -891 */ -892 CompletableFutureString getProcedures(); -893 -894 /** -895 * List locks. -896 * @return lock list JSON wrapped by {@link CompletableFuture} -897 */ -898 CompletableFutureString getLocks(); -899 -900 /** -901 * Mark region server(s) as decommissioned to prevent additional regions from getting -902 * assigned to them. Optionally unload the regions on the servers. If there are multiple servers -903 * to be decommissioned, decommissioning them at the same time can prevent wasteful region -904 * movements. Region unloading is asynchronous. -905 * @param servers The list of servers to decommission. -906 * @param offload True to offload the regions from the decommissioned servers -907 */ -908 CompletableFutureVoid decommissionRegionServers(ListServerName servers, boolean offload); -909 -910 /** -911 * List region servers marked as decommissioned, which can not be assigned regions. -912 * @return List of decommissioned region servers wrapped by {@link CompletableFuture} -913 */ -914 CompletableFutureListServerName listDecommissionedRegionServers(); -915 -916 /** -917 * Remove decommission marker from a region server to allow regions assignments. Load regions onto -918 * the server if a list of regions is given. Region loading is asynchronous. -919 * @param server The server to recommission. -920 * @param encodedRegionNames Regions to load onto the server. -921 */ -922 CompletableFutureVoid recommissionRegionServer(ServerName server, -923 Listbyte[] encodedRegionNames); -924 -925 /** -926 * @return cluster status wrapped by {@link CompletableFuture} -927 */ -928 CompletableFutureClusterMetrics getClusterMetrics(); -929 -930 /** -931 * @return cluster status wrapped by {@link CompletableFuture} -932 */ -933 CompletableFutureClusterMetrics getClusterMetrics(EnumSetOption options); -934 -935 /** -936 * @return current master server name wrapped by {@link CompletableFuture} -937 */ -938 default CompletableFutureServerName getMaster() { -939return getClusterMetrics(EnumSet.of(Option.MASTER)).thenApply(ClusterMetrics::getMasterName); -940 } -941 -942 /** -943 * @return current backup master list wrapped by {@link CompletableFuture} -944 */ -945 default CompletableFutureCollectionServerName getBackupMasters() { -946return getClusterMetrics(EnumSet.of(Option.BACKUP_MASTERS)) -947 .thenApply(ClusterMetrics::getBackupMasterNames); -948 } -949 -950 /** -951 * @return current live region servers list wrapped by {@link CompletableFuture} -952 */ -953 default CompletableFutureCollectionServerName getRegionServers() { -954return getClusterMetrics(EnumSet.of(Option.LIVE_SERVERS)) -955 .thenApply(cm - cm.getLiveServerMetrics().keySet()); -956 } -957 -958 /** -959 * @return a list of master coprocessors wrapped by {@link CompletableFuture} -960 */ -961 default CompletableFutureListString getMasterCoprocessorNames() { -962return getClusterMetrics(EnumSet.of(Option.MASTER_COPROCESSORS)) -963 .thenApply(ClusterMetrics::getMasterCoprocessorNames); -964 } -965 -966 /** -967 * Get the info port of the current master if one is available. -968 * @return master info port -969 */ -970 default CompletableFutureInteger getMasterInfoPort() { -971return getClusterMetrics(EnumSet.of(Option.MASTER_INFO_PORT)).thenApply( -972 ClusterMetrics::getMasterInfoPort); -973 } -974 -975 /** -976 * Shuts down the HBase cluster. -977 */ -978 CompletableFutureVoid shutdown(); -979 -980 /** -981 * Shuts down the current HBase master only. -982 */ -983 CompletableFutureVoid stopMaster(); -984 -985 /** -986 * Stop the designated regionserver. -987 * @param serverName -988 */ -989 CompletableFutureVoid
[21/51] [partial] hbase-site git commit: Published site at 8eaaa63114a64bcaeaf0ed9bdd88615ee22255c1.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f6f9d4f3/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html index 0cf012a..976894f 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html @@ -63,3884 +63,3883 @@ 055import javax.servlet.http.HttpServlet; 056import javax.servlet.http.HttpServletRequest; 057import javax.servlet.http.HttpServletResponse; -058 -059import org.apache.commons.lang3.StringUtils; -060import org.apache.hadoop.conf.Configuration; -061import org.apache.hadoop.fs.Path; -062import org.apache.hadoop.hbase.ChoreService; -063import org.apache.hadoop.hbase.ClusterId; -064import org.apache.hadoop.hbase.ClusterMetrics; -065import org.apache.hadoop.hbase.ClusterMetrics.Option; -066import org.apache.hadoop.hbase.ClusterMetricsBuilder; -067import org.apache.hadoop.hbase.CompoundConfiguration; -068import org.apache.hadoop.hbase.DoNotRetryIOException; -069import org.apache.hadoop.hbase.HBaseIOException; -070import org.apache.hadoop.hbase.HBaseInterfaceAudience; -071import org.apache.hadoop.hbase.HConstants; -072import org.apache.hadoop.hbase.InvalidFamilyOperationException; -073import org.apache.hadoop.hbase.MasterNotRunningException; -074import org.apache.hadoop.hbase.MetaTableAccessor; -075import org.apache.hadoop.hbase.NamespaceDescriptor; -076import org.apache.hadoop.hbase.PleaseHoldException; -077import org.apache.hadoop.hbase.ReplicationPeerNotFoundException; -078import org.apache.hadoop.hbase.ServerName; -079import org.apache.hadoop.hbase.TableDescriptors; -080import org.apache.hadoop.hbase.TableName; -081import org.apache.hadoop.hbase.TableNotDisabledException; -082import org.apache.hadoop.hbase.TableNotFoundException; -083import org.apache.hadoop.hbase.UnknownRegionException; -084import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor; -085import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder; -086import org.apache.hadoop.hbase.client.MasterSwitchType; -087import org.apache.hadoop.hbase.client.RegionInfo; -088import org.apache.hadoop.hbase.client.RegionInfoBuilder; -089import org.apache.hadoop.hbase.client.Result; -090import org.apache.hadoop.hbase.client.TableDescriptor; -091import org.apache.hadoop.hbase.client.TableDescriptorBuilder; -092import org.apache.hadoop.hbase.client.TableState; -093import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; -094import org.apache.hadoop.hbase.exceptions.DeserializationException; -095import org.apache.hadoop.hbase.exceptions.MergeRegionException; -096import org.apache.hadoop.hbase.executor.ExecutorType; -097import org.apache.hadoop.hbase.favored.FavoredNodesManager; -098import org.apache.hadoop.hbase.favored.FavoredNodesPromoter; -099import org.apache.hadoop.hbase.http.InfoServer; -100import org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils; -101import org.apache.hadoop.hbase.ipc.RpcServer; -102import org.apache.hadoop.hbase.ipc.ServerNotRunningYetException; -103import org.apache.hadoop.hbase.log.HBaseMarkers; -104import org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode; -105import org.apache.hadoop.hbase.master.assignment.AssignProcedure; -106import org.apache.hadoop.hbase.master.assignment.AssignmentManager; -107import org.apache.hadoop.hbase.master.assignment.MergeTableRegionsProcedure; -108import org.apache.hadoop.hbase.master.assignment.MoveRegionProcedure; -109import org.apache.hadoop.hbase.master.assignment.RegionStateNode; -110import org.apache.hadoop.hbase.master.assignment.RegionStates; -111import org.apache.hadoop.hbase.master.assignment.TransitRegionStateProcedure; -112import org.apache.hadoop.hbase.master.assignment.UnassignProcedure; -113import org.apache.hadoop.hbase.master.balancer.BalancerChore; -114import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer; -115import org.apache.hadoop.hbase.master.balancer.ClusterStatusChore; -116import org.apache.hadoop.hbase.master.balancer.LoadBalancerFactory; -117import org.apache.hadoop.hbase.master.cleaner.CleanerChore; -118import org.apache.hadoop.hbase.master.cleaner.HFileCleaner; -119import org.apache.hadoop.hbase.master.cleaner.LogCleaner; -120import org.apache.hadoop.hbase.master.cleaner.ReplicationBarrierCleaner; -121import org.apache.hadoop.hbase.master.locking.LockManager; -122import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan; -123import org.apache.hadoop.hbase.master.normalizer.NormalizationPlan.PlanType; -124import org.apache.hadoop.hbase.master.normalizer.RegionNormalizer; -125import
[21/51] [partial] hbase-site git commit: Published site at cd161d976ef47b84e904f2d54bac65d2f3417c2a.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fa1bebf8/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ThrowableAbortable.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ThrowableAbortable.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ThrowableAbortable.html index a5789e0..93a57cb 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ThrowableAbortable.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ThrowableAbortable.html @@ -238,4120 +238,4119 @@ 230 * @see Admin 231 */ 232@InterfaceAudience.Private -233@InterfaceStability.Evolving -234public class HBaseAdmin implements Admin { -235 private static final Logger LOG = LoggerFactory.getLogger(HBaseAdmin.class); -236 -237 private ClusterConnection connection; -238 -239 private final Configuration conf; -240 private final long pause; -241 private final int numRetries; -242 private final int syncWaitTimeout; -243 private boolean aborted; -244 private int operationTimeout; -245 private int rpcTimeout; -246 -247 private RpcRetryingCallerFactory rpcCallerFactory; -248 private RpcControllerFactory rpcControllerFactory; -249 -250 private NonceGenerator ng; -251 -252 @Override -253 public int getOperationTimeout() { -254return operationTimeout; -255 } -256 -257 HBaseAdmin(ClusterConnection connection) throws IOException { -258this.conf = connection.getConfiguration(); -259this.connection = connection; -260 -261// TODO: receive ConnectionConfiguration here rather than re-parsing these configs every time. -262this.pause = this.conf.getLong(HConstants.HBASE_CLIENT_PAUSE, -263 HConstants.DEFAULT_HBASE_CLIENT_PAUSE); -264this.numRetries = this.conf.getInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, -265 HConstants.DEFAULT_HBASE_CLIENT_RETRIES_NUMBER); -266this.operationTimeout = this.conf.getInt(HConstants.HBASE_CLIENT_OPERATION_TIMEOUT, -267 HConstants.DEFAULT_HBASE_CLIENT_OPERATION_TIMEOUT); -268this.rpcTimeout = this.conf.getInt(HConstants.HBASE_RPC_TIMEOUT_KEY, -269 HConstants.DEFAULT_HBASE_RPC_TIMEOUT); -270this.syncWaitTimeout = this.conf.getInt( -271 "hbase.client.sync.wait.timeout.msec", 10 * 6); // 10min -272 -273this.rpcCallerFactory = connection.getRpcRetryingCallerFactory(); -274this.rpcControllerFactory = connection.getRpcControllerFactory(); -275 -276this.ng = this.connection.getNonceGenerator(); -277 } -278 -279 @Override -280 public void abort(String why, Throwable e) { -281// Currently does nothing but throw the passed message and exception -282this.aborted = true; -283throw new RuntimeException(why, e); -284 } -285 -286 @Override -287 public boolean isAborted() { -288return this.aborted; -289 } -290 -291 @Override -292 public boolean abortProcedure(final long procId, final boolean mayInterruptIfRunning) -293 throws IOException { -294return get(abortProcedureAsync(procId, mayInterruptIfRunning), this.syncWaitTimeout, -295 TimeUnit.MILLISECONDS); -296 } -297 -298 @Override -299 public FutureBoolean abortProcedureAsync(final long procId, final boolean mayInterruptIfRunning) -300 throws IOException { -301Boolean abortProcResponse = -302executeCallable(new MasterCallableAbortProcedureResponse(getConnection(), -303getRpcControllerFactory()) { -304 @Override -305 protected AbortProcedureResponse rpcCall() throws Exception { -306AbortProcedureRequest abortProcRequest = -307 AbortProcedureRequest.newBuilder().setProcId(procId).build(); -308return master.abortProcedure(getRpcController(), abortProcRequest); -309 } -310}).getIsProcedureAborted(); -311return new AbortProcedureFuture(this, procId, abortProcResponse); -312 } -313 -314 @Override -315 public ListTableDescriptor listTableDescriptors() throws IOException { -316return listTableDescriptors((Pattern)null, false); -317 } -318 -319 @Override -320 public ListTableDescriptor listTableDescriptors(Pattern pattern) throws IOException { -321return listTableDescriptors(pattern, false); -322 } -323 -324 @Override -325 public ListTableDescriptor listTableDescriptors(Pattern pattern, boolean includeSysTables) -326 throws IOException { -327return executeCallable(new MasterCallableListTableDescriptor(getConnection(), -328getRpcControllerFactory()) { -329 @Override -330 protected ListTableDescriptor rpcCall() throws Exception { -331GetTableDescriptorsRequest req = -332 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables); -333return ProtobufUtil.toTableDescriptorList(master.getTableDescriptors(getRpcController(), -334req)); -335 } -336}); -337 } -338 -339
[21/51] [partial] hbase-site git commit: Published site at c6a65ba63fce85ac7c4b62b96ef2bbe6c35d2f00.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/293abb17/devapidocs/src-html/org/apache/hadoop/hbase/filter/ValueFilter.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/filter/ValueFilter.html b/devapidocs/src-html/org/apache/hadoop/hbase/filter/ValueFilter.html index 2f19834..9b23394 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/filter/ValueFilter.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/filter/ValueFilter.html @@ -29,140 +29,135 @@ 021 022import java.io.IOException; 023import java.util.ArrayList; -024import java.util.Objects; -025 -026import org.apache.hadoop.hbase.Cell; -027import org.apache.hadoop.hbase.CompareOperator; -028import org.apache.yetus.audience.InterfaceAudience; -029import org.apache.hadoop.hbase.exceptions.DeserializationException; -030import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -031import org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos; -032import org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException; -033 -034/** -035 * This filter is used to filter based on column value. It takes an -036 * operator (equal, greater, not equal, etc) and a byte [] comparator for the -037 * cell value. -038 * p -039 * This filter can be wrapped with {@link WhileMatchFilter} and {@link SkipFilter} -040 * to add more control. -041 * p -042 * Multiple filters can be combined using {@link FilterList}. -043 * p -044 * To test the value of a single qualifier when scanning multiple qualifiers, -045 * use {@link SingleColumnValueFilter}. -046 */ -047@InterfaceAudience.Public -048public class ValueFilter extends CompareFilter { -049 -050 /** -051 * Constructor. -052 * @param valueCompareOp the compare op for value matching -053 * @param valueComparator the comparator for value matching -054 * @deprecated Since 2.0.0. Will be removed in 3.0.0. -055 * Use {@link #ValueFilter(CompareOperator, ByteArrayComparable)} -056 */ -057 public ValueFilter(final CompareOp valueCompareOp, -058 final ByteArrayComparable valueComparator) { -059super(valueCompareOp, valueComparator); -060 } -061 -062 /** -063 * Constructor. -064 * @param valueCompareOp the compare op for value matching -065 * @param valueComparator the comparator for value matching -066 */ -067 public ValueFilter(final CompareOperator valueCompareOp, -068 final ByteArrayComparable valueComparator) { -069super(valueCompareOp, valueComparator); -070 } -071 -072 @Deprecated -073 @Override -074 public ReturnCode filterKeyValue(final Cell c) { -075return filterCell(c); -076 } -077 -078 @Override -079 public ReturnCode filterCell(final Cell c) { -080if (compareValue(getCompareOperator(), this.comparator, c)) { -081 return ReturnCode.SKIP; -082} -083return ReturnCode.INCLUDE; -084 } -085 -086 public static Filter createFilterFromArguments(ArrayListbyte [] filterArguments) { -087@SuppressWarnings("rawtypes") // for arguments -088ArrayList arguments = CompareFilter.extractArguments(filterArguments); -089CompareOperator compareOp = (CompareOperator)arguments.get(0); -090ByteArrayComparable comparator = (ByteArrayComparable)arguments.get(1); -091return new ValueFilter(compareOp, comparator); -092 } -093 -094 /** -095 * @return The filter serialized using pb -096 */ -097 @Override -098 public byte [] toByteArray() { -099FilterProtos.ValueFilter.Builder builder = -100 FilterProtos.ValueFilter.newBuilder(); -101 builder.setCompareFilter(super.convert()); -102return builder.build().toByteArray(); -103 } -104 -105 /** -106 * @param pbBytes A pb serialized {@link ValueFilter} instance -107 * @return An instance of {@link ValueFilter} made from codebytes/code -108 * @throws DeserializationException -109 * @see #toByteArray -110 */ -111 public static ValueFilter parseFrom(final byte [] pbBytes) -112 throws DeserializationException { -113FilterProtos.ValueFilter proto; -114try { -115 proto = FilterProtos.ValueFilter.parseFrom(pbBytes); -116} catch (InvalidProtocolBufferException e) { -117 throw new DeserializationException(e); -118} -119final CompareOperator valueCompareOp = -120 CompareOperator.valueOf(proto.getCompareFilter().getCompareOp().name()); -121ByteArrayComparable valueComparator = null; -122try { -123 if (proto.getCompareFilter().hasComparator()) { -124valueComparator = ProtobufUtil.toComparator(proto.getCompareFilter().getComparator()); -125 } -126} catch (IOException ioe) { -127 throw new DeserializationException(ioe); -128} -129return new ValueFilter(valueCompareOp,valueComparator); -130 } -131 -132 /** -133 * @return true if and only if the fields of the filter that are serialized -134 * are equal to the
[21/51] [partial] hbase-site git commit: Published site at 7c1fad4992a169a35b4457e6f4afcb30d04406e9.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/74f60271/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTableWrapperAggregate.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTableWrapperAggregate.html b/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTableWrapperAggregate.html index 92ec202..99b030e 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTableWrapperAggregate.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsTableWrapperAggregate.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":6}; +var methods = {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -129,44 +129,90 @@ public interface long +getAvgRegionSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) +Get the average region size for this table + + + +long +getAvgStoreFileAge(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) + + +long getCpRequestsCount(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) Get the number of CoprocessorService requests that have been issued against this table - + +long +getFilteredReadRequestCount(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) +Get the total number of filtered read requests that have been issued against this table + + + long -getMemStoresSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) +getMaxStoreFileAge(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) + + +long +getMemStoreSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) Get the memory store size against this table - + +long +getMinStoreFileAge(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) + + +long +getNumReferenceFiles(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) + + +long +getNumRegions(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) +Get the number of regions hosted on for this table + + + long -getReadRequestsCount(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) +getNumStoreFiles(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) +Get the number of store files hosted for this table + + + +long +getNumStores(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) +Get the number of stores hosted on for this table + + + +long +getReadRequestCount(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) Get the number of read requests that have been issued against this table - + long -getStoreFilesSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) +getStoreFileSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) Get the store file size against this table - + long getTableSize(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) Get the table region size against this table - + long getTotalRequestsCount(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) -Get the total number of requests that have been issued against this table +Get the total number of requests that have been issued for this table - + long -getWriteRequestsCount(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringtable) -Get the number of write requests that have been issued
[21/51] [partial] hbase-site git commit: Published site at 3afe9fb7e6ebfa71187cbe131558a83fae61cecd.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/424d7e41/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html index bd7445a..3504442 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HStore.html @@ -1720,459 +1720,459 @@ 1712 LOG.error("Error trying to determine if store has references, assuming references exists", 1713ioe); 1714 return true; -1715} -1716 } -1717 -1718 /** -1719 * getter for CompactionProgress object -1720 * @return CompactionProgress object; can be null -1721 */ -1722 public CompactionProgress getCompactionProgress() { -1723return this.storeEngine.getCompactor().getProgress(); -1724 } -1725 -1726 @Override -1727 public boolean shouldPerformMajorCompaction() throws IOException { -1728for (HStoreFile sf : this.storeEngine.getStoreFileManager().getStorefiles()) { -1729 // TODO: what are these reader checks all over the place? -1730 if (sf.getReader() == null) { -1731LOG.debug("StoreFile {} has null Reader", sf); -1732return false; -1733 } -1734} -1735return storeEngine.getCompactionPolicy().shouldPerformMajorCompaction( -1736 this.storeEngine.getStoreFileManager().getStorefiles()); -1737 } -1738 -1739 public OptionalCompactionContext requestCompaction() throws IOException { -1740return requestCompaction(NO_PRIORITY, CompactionLifeCycleTracker.DUMMY, null); -1741 } -1742 -1743 public OptionalCompactionContext requestCompaction(int priority, -1744 CompactionLifeCycleTracker tracker, User user) throws IOException { -1745// don't even select for compaction if writes are disabled -1746if (!this.areWritesEnabled()) { -1747 return Optional.empty(); -1748} -1749// Before we do compaction, try to get rid of unneeded files to simplify things. -1750removeUnneededFiles(); -1751 -1752final CompactionContext compaction = storeEngine.createCompaction(); -1753CompactionRequestImpl request = null; -1754this.lock.readLock().lock(); -1755try { -1756 synchronized (filesCompacting) { -1757// First, see if coprocessor would want to override selection. -1758if (this.getCoprocessorHost() != null) { -1759 final ListHStoreFile candidatesForCoproc = compaction.preSelect(this.filesCompacting); -1760 boolean override = getCoprocessorHost().preCompactSelection(this, -1761 candidatesForCoproc, tracker, user); -1762 if (override) { -1763// Coprocessor is overriding normal file selection. -1764compaction.forceSelect(new CompactionRequestImpl(candidatesForCoproc)); -1765 } -1766} -1767 -1768// Normal case - coprocessor is not overriding file selection. -1769if (!compaction.hasSelection()) { -1770 boolean isUserCompaction = priority == Store.PRIORITY_USER; -1771 boolean mayUseOffPeak = offPeakHours.isOffPeakHour() -1772 offPeakCompactionTracker.compareAndSet(false, true); -1773 try { -1774 compaction.select(this.filesCompacting, isUserCompaction, -1775 mayUseOffPeak, forceMajor filesCompacting.isEmpty()); -1776 } catch (IOException e) { -1777if (mayUseOffPeak) { -1778 offPeakCompactionTracker.set(false); -1779} -1780throw e; -1781 } -1782 assert compaction.hasSelection(); -1783 if (mayUseOffPeak !compaction.getRequest().isOffPeak()) { -1784// Compaction policy doesn't want to take advantage of off-peak. -1785 offPeakCompactionTracker.set(false); -1786 } -1787} -1788if (this.getCoprocessorHost() != null) { -1789 this.getCoprocessorHost().postCompactSelection( -1790 this, ImmutableList.copyOf(compaction.getRequest().getFiles()), tracker, -1791 compaction.getRequest(), user); -1792} -1793// Finally, we have the resulting files list. Check if we have any files at all. -1794request = compaction.getRequest(); -1795CollectionHStoreFile selectedFiles = request.getFiles(); -1796if (selectedFiles.isEmpty()) { -1797 return Optional.empty(); +1715} finally { +1716 if (reloadedStoreFiles != null) { +1717for (HStoreFile storeFile : reloadedStoreFiles) { +1718 try { +1719 storeFile.closeStoreFile(false); +1720 } catch (IOException ioe) { +1721LOG.warn("Encountered exception closing " + storeFile + ": " + ioe.getMessage()); +1722// continue with
[21/51] [partial] hbase-site git commit: Published site at a452487a9b82bfd33bc10683c3f8b8ae74d58883.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cf79db0/devapidocs/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.html b/devapidocs/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.html index ac54f87..368fdb4 100644 --- a/devapidocs/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.html +++ b/devapidocs/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":9,"i2":10,"i3":42,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":9,"i10":10,"i11":10,"i12":10}; +var methods = {"i0":10,"i1":9,"i2":10,"i3":10,"i4":42,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":9,"i12":10,"i13":10,"i14":10}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"],32:["t6","Deprecated Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -120,7 +120,7 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Public -public class ColumnPaginationFilter +public class ColumnPaginationFilter extends FilterBase A filter, based on the ColumnCountGetFilter, takes two arguments: limit and offset. This filter can be used for row-based indexing, where references to other tables are stored across many columns, @@ -234,60 +234,68 @@ extends createFilterFromArguments(https://docs.oracle.com/javase/8/docs/api/java/util/ArrayList.html?is-external=true; title="class or interface in java.util">ArrayListbyte[]filterArguments) +boolean +equals(https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Objectobj) + + Filter.ReturnCode filterCell(Cellc) A way to filter based on the column family, column qualifier and/or the column value. - + Filter.ReturnCode filterKeyValue(Cellc) Deprecated. - + boolean filterRowKey(Cellcell) Filters a row based on the row key. - + byte[] getColumnOffset() - + int getLimit() - + Cell getNextCellHint(Cellcell) Filters that are not sure which key must be next seeked to, can inherit this implementation that, by default, returns a null Cell. - + int getOffset() - + +int +hashCode() + + static ColumnPaginationFilter parseFrom(byte[]pbBytes) - + void reset() Filters that are purely stateless and do nothing in their reset() methods can inherit this null/empty implementation. - + byte[] toByteArray() Return length 0 byte array for Filters that don't require special serialization - + https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String toString() Return filter's info for debugging and logging purpose. @@ -313,7 +321,7 @@ extends Methods inherited from classjava.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object -https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--; title="class or interface in java.lang">clone, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-; title="class or interface in java.lang">equals, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--; title="class or interface in java.lang">finalize, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--; title="class or interface in java.lang">getClass, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--; title="class or interface in java.lang">hashCode, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--; title="class or interface in java.lang">notify, https://docs.oracle.com/javase/8/docs/api/ja va/lang/Object.html?is-external=true#notifyAll--" title="class or interface in java.lang">notifyAll, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait--; title="class or interface in java.lang">wait, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-; title="class or interface in java.lang">wait, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-int-; title="class or interface in java.lang">wait +https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--; title="class or interface in java.lang">clone, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--; title="class or interface in java.lang">finalize, https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--; title="class or interface in java.lang">getClass,
[21/51] [partial] hbase-site git commit: Published site at 6a5b4f2a5c188f8eef4f2250b8b7db7dd1e750e4.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1ff05a18/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html index db8431b..a8cb7c4 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html @@ -885,7766 +885,7797 @@ 877 * @return What the next sequence (edit) id should be. 878 * @throws IOException e 879 */ -880 private long initialize(final CancelableProgressable reporter) throws IOException { -881 -882//Refuse to open the region if there is no column family in the table -883if (htableDescriptor.getColumnFamilyCount() == 0) { -884 throw new DoNotRetryIOException("Table " + htableDescriptor.getTableName().getNameAsString()+ -885 " should have at least one column family."); -886} -887 -888MonitoredTask status = TaskMonitor.get().createStatus("Initializing region " + this); -889long nextSeqId = -1; -890try { -891 nextSeqId = initializeRegionInternals(reporter, status); -892 return nextSeqId; -893} finally { -894 // nextSeqid will be -1 if the initialization fails. -895 // At least it will be 0 otherwise. -896 if (nextSeqId == -1) { -897status.abort("Exception during region " + getRegionInfo().getRegionNameAsString() + -898 " initialization."); -899 } -900} -901 } -902 -903 private long initializeRegionInternals(final CancelableProgressable reporter, -904 final MonitoredTask status) throws IOException { -905if (coprocessorHost != null) { -906 status.setStatus("Running coprocessor pre-open hook"); -907 coprocessorHost.preOpen(); -908} -909 -910// Write HRI to a file in case we need to recover hbase:meta -911// Only the primary replica should write .regioninfo -912if (this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) { -913 status.setStatus("Writing region info on filesystem"); -914 fs.checkRegionInfoOnFilesystem(); -915} -916 -917// Initialize all the HStores -918status.setStatus("Initializing all the Stores"); -919long maxSeqId = initializeStores(reporter, status); -920this.mvcc.advanceTo(maxSeqId); -921if (ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) { -922 CollectionHStore stores = this.stores.values(); -923 try { -924// update the stores that we are replaying -925LOG.debug("replaying wal for " + this.getRegionInfo().getEncodedName()); -926 stores.forEach(HStore::startReplayingFromWAL); -927// Recover any edits if available. -928maxSeqId = Math.max(maxSeqId, -929 replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter, status)); -930// Make sure mvcc is up to max. -931this.mvcc.advanceTo(maxSeqId); -932 } finally { -933LOG.debug("stopping wal replay for " + this.getRegionInfo().getEncodedName()); -934// update the stores that we are done replaying -935 stores.forEach(HStore::stopReplayingFromWAL); -936 } -937} -938this.lastReplayedOpenRegionSeqId = maxSeqId; -939 -940 this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this)); -941this.writestate.flushRequested = false; -942this.writestate.compacting.set(0); -943 -944if (this.writestate.writesEnabled) { -945 LOG.debug("Cleaning up temporary data for " + this.getRegionInfo().getEncodedName()); -946 // Remove temporary data left over from old regions -947 status.setStatus("Cleaning up temporary data from old regions"); -948 fs.cleanupTempDir(); -949} -950 -951if (this.writestate.writesEnabled) { -952 status.setStatus("Cleaning up detritus from prior splits"); -953 // Get rid of any splits or merges that were lost in-progress. Clean out -954 // these directories here on open. We may be opening a region that was -955 // being split but we crashed in the middle of it all. -956 LOG.debug("Cleaning up detritus for " + this.getRegionInfo().getEncodedName()); -957 fs.cleanupAnySplitDetritus(); -958 fs.cleanupMergesDir(); -959} +880 @VisibleForTesting +881 long initialize(final CancelableProgressable reporter) throws IOException { +882 +883//Refuse to open the region if there is no column family in the table +884if (htableDescriptor.getColumnFamilyCount() == 0) { +885 throw new DoNotRetryIOException("Table " + htableDescriptor.getTableName().getNameAsString()+ +886 " should have at least one column family."); +887} +888
[21/51] [partial] hbase-site git commit: Published site at 63f2d3cbdc8151f5f61f33e0a078c51b9ac076a5.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7ae6a80c/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStateNode.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStateNode.html b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStateNode.html new file mode 100644 index 000..e31a01d --- /dev/null +++ b/devapidocs/org/apache/hadoop/hbase/master/assignment/class-use/RegionStateNode.html @@ -0,0 +1,607 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +Uses of Class org.apache.hadoop.hbase.master.assignment.RegionStateNode (Apache HBase 3.0.0-SNAPSHOT API) + + + + + + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +Prev +Next + + +Frames +NoFrames + + +AllClasses + + + + + + + + + + +Uses of Classorg.apache.hadoop.hbase.master.assignment.RegionStateNode + + + + + +Packages that use RegionStateNode + +Package +Description + + + +org.apache.hadoop.hbase.master.assignment + + + + + + + + + + +Uses of RegionStateNode in org.apache.hadoop.hbase.master.assignment + +Fields in org.apache.hadoop.hbase.master.assignment declared as RegionStateNode + +Modifier and Type +Field and Description + + + +private RegionStateNode +RegionStates.RegionFailedOpen.regionNode + + + + +Fields in org.apache.hadoop.hbase.master.assignment with type parameters of type RegionStateNode + +Modifier and Type +Field and Description + + + +private https://docs.oracle.com/javase/8/docs/api/java/util/ArrayList.html?is-external=true; title="class or interface in java.util">ArrayListRegionStateNode +AssignmentManager.pendingAssignQueue + + +private https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentSkipListMap.html?is-external=true; title="class or interface in java.util.concurrent">ConcurrentSkipListMapRegionInfo,RegionStateNode +RegionStates.regionInTransition + + +private https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentSkipListMap.html?is-external=true; title="class or interface in java.util.concurrent">ConcurrentSkipListMapRegionInfo,RegionStateNode +RegionStates.regionOffline +Regions marked as offline on a read of hbase:meta. + + + +private https://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">SetRegionStateNode +ServerStateNode.regions + + +private https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentSkipListMap.html?is-external=true; title="class or interface in java.util.concurrent">ConcurrentSkipListMapbyte[],RegionStateNode +RegionStates.regionsMap +RegionName -- i.e. + + + +private https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true; title="class or interface in java.util.concurrent">ConcurrentMapRegionInfo,RegionStateNode +RegionStateNode.ritMap + + + + +Methods in org.apache.hadoop.hbase.master.assignment that return RegionStateNode + +Modifier and Type +Method and Description + + + +(package private) RegionStateNode +RegionStates.createRegionStateNode(RegionInforegionInfo) + + +RegionStateNode +RegionStates.getOrCreateRegionStateNode(RegionInforegionInfo) + + +RegionStateNode +RegionTransitionProcedure.getRegionState(MasterProcedureEnvenv) +Deprecated. + + + +RegionStateNode +RegionStates.RegionFailedOpen.getRegionStateNode() + + +private RegionStateNode +TransitRegionStateProcedure.getRegionStateNode(MasterProcedureEnvenv) + + +RegionStateNode +RegionStates.getRegionStateNode(RegionInforegionInfo) + + +(package private) RegionStateNode +RegionStates.getRegionStateNodeFromName(byte[]regionName) + + + + +Methods in org.apache.hadoop.hbase.master.assignment that return types with arguments of type RegionStateNode + +Modifier and Type +Method and Description + + + +https://docs.oracle.com/javase/8/docs/api/java/util/Set.html?is-external=true; title="class or interface in java.util">SetRegionStateNode +ServerStateNode.getRegions() + + +https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionStateNode +RegionStates.getRegionsInTransition() + + +https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionStateNode
[21/51] [partial] hbase-site git commit: Published site at 092efb42749bf7fc6ad338c96aae8e7b9d3a2c74.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f3d62514/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.html b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.html index 9501e97..a10ddfe 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/mapreduce/WALPlayer.html @@ -131,277 +131,279 @@ 123 } 124} 125 } catch (InterruptedException e) { -126e.printStackTrace(); -127 } -128} -129 -130@Override -131public void setup(Context context) throws IOException { -132 Configuration conf = context.getConfiguration(); -133 String[] tables = conf.getStrings(TABLES_KEY); -134 this.multiTableSupport = conf.getBoolean(MULTI_TABLES_SUPPORT, false); -135 for (String table : tables) { -136tableSet.add(table); -137 } -138} -139 } -140 -141 /** -142 * A mapper that writes out {@link Mutation} to be directly applied to a running HBase instance. -143 */ -144 protected static class WALMapper -145 extends MapperWALKey, WALEdit, ImmutableBytesWritable, Mutation { -146private MapTableName, TableName tables = new TreeMap(); -147 -148@Override -149public void map(WALKey key, WALEdit value, Context context) throws IOException { -150 try { -151if (tables.isEmpty() || tables.containsKey(key.getTableName())) { -152 TableName targetTable = -153 tables.isEmpty() ? key.getTableName() : tables.get(key.getTableName()); -154 ImmutableBytesWritable tableOut = new ImmutableBytesWritable(targetTable.getName()); -155 Put put = null; -156 Delete del = null; -157 Cell lastCell = null; -158 for (Cell cell : value.getCells()) { -159// filtering WAL meta entries -160if (WALEdit.isMetaEditFamily(cell)) { -161 continue; -162} -163 -164// Allow a subclass filter out this cell. -165if (filter(context, cell)) { -166 // A WALEdit may contain multiple operations (HBASE-3584) and/or -167 // multiple rows (HBASE-5229). -168 // Aggregate as much as possible into a single Put/Delete -169 // operation before writing to the context. -170 if (lastCell == null || lastCell.getTypeByte() != cell.getTypeByte() -171 || !CellUtil.matchingRows(lastCell, cell)) { -172// row or type changed, write out aggregate KVs. -173if (put != null) { -174 context.write(tableOut, put); -175} -176if (del != null) { -177 context.write(tableOut, del); -178} -179if (CellUtil.isDelete(cell)) { -180 del = new Delete(CellUtil.cloneRow(cell)); -181} else { -182 put = new Put(CellUtil.cloneRow(cell)); -183} -184 } -185 if (CellUtil.isDelete(cell)) { -186del.add(cell); -187 } else { -188put.add(cell); -189 } -190} -191lastCell = cell; -192 } -193 // write residual KVs -194 if (put != null) { -195context.write(tableOut, put); -196 } -197 if (del != null) { -198context.write(tableOut, del); -199 } -200} -201 } catch (InterruptedException e) { -202e.printStackTrace(); -203 } -204} -205 -206protected boolean filter(Context context, final Cell cell) { -207 return true; -208} -209 -210@Override -211protected void -212cleanup(MapperWALKey, WALEdit, ImmutableBytesWritable, Mutation.Context context) -213throws IOException, InterruptedException { -214 super.cleanup(context); -215} -216 -217@Override -218public void setup(Context context) throws IOException { -219 String[] tableMap = context.getConfiguration().getStrings(TABLE_MAP_KEY); -220 String[] tablesToUse = context.getConfiguration().getStrings(TABLES_KEY); -221 if (tableMap == null) { -222tableMap = tablesToUse; -223 } -224 if (tablesToUse == null) { -225// Then user wants all tables. -226 } else if (tablesToUse.length != tableMap.length) { -227// this can only happen when WALMapper is used directly by a class other than WALPlayer -228throw new IOException("Incorrect table mapping specified ."); -229 } -230 int i = 0; -231 if (tablesToUse != null) { -232for (String table : tablesToUse) { -233 tables.put(TableName.valueOf(table),
[21/51] [partial] hbase-site git commit: Published site at 613d831429960348dc42c3bdb6ea5d31be15c81c.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7cf6034b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.html b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.html index b7b4236..3d1edb3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/HFileBlock.html @@ -259,1863 +259,1867 @@ 251 * + Metadata! + = See note on BLOCK_METADATA_SPACE above. 252 * ++ 253 * /code -254 * @see #serialize(ByteBuffer) +254 * @see #serialize(ByteBuffer, boolean) 255 */ -256 static final CacheableDeserializerCacheable BLOCK_DESERIALIZER = -257 new CacheableDeserializerCacheable() { -258@Override -259public HFileBlock deserialize(ByteBuff buf, boolean reuse, MemoryType memType) -260throws IOException { -261 // The buf has the file block followed by block metadata. -262 // Set limit to just before the BLOCK_METADATA_SPACE then rewind. -263 buf.limit(buf.limit() - BLOCK_METADATA_SPACE).rewind(); -264 // Get a new buffer to pass the HFileBlock for it to 'own'. -265 ByteBuff newByteBuff; -266 if (reuse) { -267newByteBuff = buf.slice(); -268 } else { -269int len = buf.limit(); -270newByteBuff = new SingleByteBuff(ByteBuffer.allocate(len)); -271newByteBuff.put(0, buf, buf.position(), len); -272 } -273 // Read out the BLOCK_METADATA_SPACE content and shove into our HFileBlock. -274 buf.position(buf.limit()); -275 buf.limit(buf.limit() + HFileBlock.BLOCK_METADATA_SPACE); -276 boolean usesChecksum = buf.get() == (byte) 1; -277 long offset = buf.getLong(); -278 int nextBlockOnDiskSize = buf.getInt(); -279 HFileBlock hFileBlock = -280 new HFileBlock(newByteBuff, usesChecksum, memType, offset, nextBlockOnDiskSize, null); -281 return hFileBlock; -282} -283 -284@Override -285public int getDeserialiserIdentifier() { -286 return DESERIALIZER_IDENTIFIER; -287} -288 -289@Override -290public HFileBlock deserialize(ByteBuff b) throws IOException { -291 // Used only in tests -292 return deserialize(b, false, MemoryType.EXCLUSIVE); -293} -294 }; -295 -296 private static final int DESERIALIZER_IDENTIFIER; -297 static { -298DESERIALIZER_IDENTIFIER = -299 CacheableDeserializerIdManager.registerDeserializer(BLOCK_DESERIALIZER); -300 } -301 -302 /** -303 * Copy constructor. Creates a shallow copy of {@code that}'s buffer. -304 */ -305 private HFileBlock(HFileBlock that) { -306this(that, false); -307 } -308 -309 /** -310 * Copy constructor. Creates a shallow/deep copy of {@code that}'s buffer as per the boolean -311 * param. -312 */ -313 private HFileBlock(HFileBlock that, boolean bufCopy) { -314init(that.blockType, that.onDiskSizeWithoutHeader, -315 that.uncompressedSizeWithoutHeader, that.prevBlockOffset, -316that.offset, that.onDiskDataSizeWithHeader, that.nextBlockOnDiskSize, that.fileContext); -317if (bufCopy) { -318 this.buf = new SingleByteBuff(ByteBuffer.wrap(that.buf.toBytes(0, that.buf.limit(; -319} else { -320 this.buf = that.buf.duplicate(); -321} -322 } -323 -324 /** -325 * Creates a new {@link HFile} block from the given fields. This constructor -326 * is used only while writing blocks and caching, -327 * and is sitting in a byte buffer and we want to stuff the block into cache. -328 * -329 * pTODO: The caller presumes no checksumming -330 * required of this block instance since going into cache; checksum already verified on -331 * underlying block data pulled in from filesystem. Is that correct? What if cache is SSD? +256 public static final CacheableDeserializerCacheable BLOCK_DESERIALIZER = new BlockDeserializer(); +257 +258 public static final class BlockDeserializer implements CacheableDeserializerCacheable { +259private BlockDeserializer() { +260} +261 +262@Override +263public HFileBlock deserialize(ByteBuff buf, boolean reuse, MemoryType memType) +264throws IOException { +265 // The buf has the file block followed by block metadata. +266 // Set limit to just before the BLOCK_METADATA_SPACE then rewind. +267 buf.limit(buf.limit() - BLOCK_METADATA_SPACE).rewind(); +268 // Get a new buffer to pass the HFileBlock for it to 'own'. +269 ByteBuff newByteBuff; +270 if (reuse) { +271newByteBuff = buf.slice(); +272 } else { +273int len = buf.limit(); +274newByteBuff = new SingleByteBuff(ByteBuffer.allocate(len)); +275newByteBuff.put(0, buf, buf.position(), len); +276 } +277 // Read out the BLOCK_METADATA_SPACE content
[21/51] [partial] hbase-site git commit: Published site at ba5d1c1f28301adc99019d9d6c4a04fac98ae511.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/804782f0/devapidocs/org/apache/hadoop/hbase/procedure2/Procedure.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/Procedure.html b/devapidocs/org/apache/hadoop/hbase/procedure2/Procedure.html index 4c6f707..19a9ffb 100644 --- a/devapidocs/org/apache/hadoop/hbase/procedure2/Procedure.html +++ b/devapidocs/org/apache/hadoop/hbase/procedure2/Procedure.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":6,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":6,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":6,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":9,"i24":10,"i25":10,"i26":9,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":9,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":6,"i52":6,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10}; +var methods = {"i0":6,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":6,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":6,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":9,"i24":10,"i25":10,"i26":9,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":9,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":6,"i54":6,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -118,60 +118,78 @@ var activeTableTab = "activeTableTab"; @InterfaceAudience.Private - @InterfaceStability.Evolving -public abstract class ProcedureTEnvironment +public abstract class ProcedureTEnvironment extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableProcedureTEnvironment -Base Procedure class responsible for Procedure Metadata; - e.g. state, submittedTime, lastUpdate, stack-indexes, etc. - - Procedures are run by a ProcedureExecutor instance. They are submitted and then - the ProcedureExecutor keeps calling execute(Object) until the Procedure is done. - Execute may be called multiple times in the case of failure or a restart, so code must be - idempotent. The return from an execute call is either: null to indicate we are done; - ourself if there is more to do; or, a set of sub-procedures that need to - be run to completion before the framework resumes our execution. - - The ProcedureExecutor keeps its - notion of Procedure State in the Procedure itself; e.g. it stamps the Procedure as INITIALIZING, - RUNNABLE, SUCCESS, etc. Here are some of the States defined in the ProcedureState enum from - protos: - - isFailed() A procedure has executed at least once and has failed. The procedure - may or may not have rolled back yet. Any procedure in FAILED state will be eventually moved - to ROLLEDBACK state. - +Base Procedure class responsible for Procedure Metadata; e.g. state, submittedTime, lastUpdate, + stack-indexes, etc. + + Procedures are run by a ProcedureExecutor instance. They are submitted and then the + ProcedureExecutor keeps calling execute(Object) until the Procedure is done. Execute may + be called multiple times in the case of failure or a restart, so code must be idempotent. The + return from an execute call is either: null to indicate we are done; ourself if there is more to + do; or, a set of sub-procedures that need to be run to completion before the framework resumes + our execution. + + The ProcedureExecutor keeps its notion of Procedure State in the Procedure itself; e.g. it stamps + the Procedure as INITIALIZING, RUNNABLE, SUCCESS, etc. Here are some of the States defined in the + ProcedureState enum from protos: + + isFailed() A procedure has executed at least once and has failed. The procedure may + or may not have rolled back yet. Any procedure in FAILED state will be eventually moved to + ROLLEDBACK state. isSuccess() A procedure is completed successfully without exception. - isFinished() As a
[21/51] [partial] hbase-site git commit: Published site at b4759ce6e72f50ccd9d410bd5917dc5a515414f1.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/df8fd1d3/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MemStoreSizeCostFunction.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MemStoreSizeCostFunction.html b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MemStoreSizeCostFunction.html index 233dba3..91b9055 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MemStoreSizeCostFunction.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/balancer/StochasticLoadBalancer.MemStoreSizeCostFunction.html @@ -540,1205 +540,1204 @@ 532 sm.getRegionMetrics().forEach((byte[] regionName, RegionMetrics rm) - { 533DequeBalancerRegionLoad rLoads = oldLoads.get(Bytes.toString(regionName)); 534if (rLoads == null) { -535 // There was nothing there -536 rLoads = new ArrayDeque(); -537} else if (rLoads.size() = numRegionLoadsToRemember) { -538 rLoads.remove(); -539} -540rLoads.add(new BalancerRegionLoad(rm)); -541 loads.put(Bytes.toString(regionName), rLoads); -542 }); -543}); -544 -545for(CostFromRegionLoadFunction cost : regionLoadFunctions) { -546 cost.setLoads(loads); -547} -548 } -549 -550 protected void initCosts(Cluster cluster) { -551for (CostFunction c:costFunctions) { -552 c.init(cluster); -553} -554 } -555 -556 protected void updateCostsWithAction(Cluster cluster, Action action) { -557for (CostFunction c : costFunctions) { -558 c.postAction(action); -559} -560 } -561 -562 /** -563 * Get the names of the cost functions -564 */ -565 public String[] getCostFunctionNames() { -566if (costFunctions == null) return null; -567String[] ret = new String[costFunctions.length]; -568for (int i = 0; i costFunctions.length; i++) { -569 CostFunction c = costFunctions[i]; -570 ret[i] = c.getClass().getSimpleName(); -571} -572 -573return ret; -574 } -575 -576 /** -577 * This is the main cost function. It will compute a cost associated with a proposed cluster -578 * state. All different costs will be combined with their multipliers to produce a double cost. -579 * -580 * @param cluster The state of the cluster -581 * @param previousCost the previous cost. This is used as an early out. -582 * @return a double of a cost associated with the proposed cluster state. This cost is an -583 * aggregate of all individual cost functions. -584 */ -585 protected double computeCost(Cluster cluster, double previousCost) { -586double total = 0; -587 -588for (int i = 0; i costFunctions.length; i++) { -589 CostFunction c = costFunctions[i]; -590 this.tempFunctionCosts[i] = 0.0; -591 -592 if (c.getMultiplier() = 0) { -593continue; -594 } -595 -596 Float multiplier = c.getMultiplier(); -597 Double cost = c.cost(); -598 -599 this.tempFunctionCosts[i] = multiplier*cost; -600 total += this.tempFunctionCosts[i]; -601 -602 if (total previousCost) { -603break; -604 } -605} -606 -607return total; -608 } -609 -610 /** Generates a candidate action to be applied to the cluster for cost function search */ -611 abstract static class CandidateGenerator { -612abstract Cluster.Action generate(Cluster cluster); -613 -614/** -615 * From a list of regions pick a random one. Null can be returned which -616 * {@link StochasticLoadBalancer#balanceCluster(Map)} recognize as signal to try a region move -617 * rather than swap. -618 * -619 * @param clusterThe state of the cluster -620 * @param server index of the server -621 * @param chanceOfNoSwap Chance that this will decide to try a move rather -622 * than a swap. -623 * @return a random {@link RegionInfo} or null if an asymmetrical move is -624 * suggested. -625 */ -626protected int pickRandomRegion(Cluster cluster, int server, double chanceOfNoSwap) { -627 // Check to see if this is just a move. -628 if (cluster.regionsPerServer[server].length == 0 || RANDOM.nextFloat() chanceOfNoSwap) { -629// signal a move only. -630return -1; -631 } -632 int rand = RANDOM.nextInt(cluster.regionsPerServer[server].length); -633 return cluster.regionsPerServer[server][rand]; -634 -635} -636protected int pickRandomServer(Cluster cluster) { -637 if (cluster.numServers 1) { -638return -1; -639 } -640 -641 return RANDOM.nextInt(cluster.numServers); -642} -643 -644protected int pickRandomRack(Cluster cluster) { -645 if (cluster.numRacks 1) { -646return -1; -647
[21/51] [partial] hbase-site git commit: Published site at e66a6603e36ecd67237ca16acd5e2de03f0d372d.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0c6f447e/apidocs/org/apache/hadoop/hbase/RegionMetrics.html -- diff --git a/apidocs/org/apache/hadoop/hbase/RegionMetrics.html b/apidocs/org/apache/hadoop/hbase/RegionMetrics.html index e4dc75c..c745b96 100644 --- a/apidocs/org/apache/hadoop/hbase/RegionMetrics.html +++ b/apidocs/org/apache/hadoop/hbase/RegionMetrics.html @@ -1,6 +1,6 @@ http://www.w3.org/TR/html4/loose.dtd;> - + @@ -19,45 +19,45 @@ } //--> var methods = {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":6,"i7":6,"i8":6,"i9":18,"i10":6,"i11":6,"i12":18,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6,"i20":6,"i21":6}; -var tabs = {65535:["t0","æææ¹æ³"],2:["t2","å®ä¾æ¹æ³"],4:["t3","æ½è±¡æ¹æ³"],16:["t5","é»è®¤æ¹æ³"]}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],16:["t5","Default Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; var tableTab = "tableTab"; var activeTableTab = "activeTableTab"; -æ¨çæµè§å¨å·²ç¦ç¨ JavaScriptã +JavaScript is disabled on your browser. -è·³è¿å¯¼èªé¾æ¥ +Skip navigation links - -æ¦è§ -ç¨åºå -ç±» -ä½¿ç¨ -æ -å·²è¿æ¶ -ç´¢å¼ -å¸®å© + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help -ä¸ä¸ä¸ªç±» -ä¸ä¸ä¸ªç±» +PrevClass +NextClass -æ¡æ¶ -æ æ¡æ¶ +Frames +NoFrames -ææç±» +AllClassesorg.apache.hadoop.hbase-æ¥å£ RegionMetrics
+Interface RegionMetrics
-
- ææå·²ç¥å®ç°ç±»:
-- RegionLoad
+- All Known Implementing Classes:
+- RegionLoad
@@ -119,15 +119,15 @@ public interface -æ¹æ³æ¦è¦
-