http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5427a45e/apidocs/org/apache/hadoop/hbase/Cell.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/Cell.html
b/apidocs/org/apache/hadoop/hbase/Cell.html
index f50a5d4..d0f0a2a 100644
--- a/apidocs/org/apache/hadoop/hbase/Cell.html
+++ b/apidocs/org/apache/hadoop/hbase/Cell.html
@@ -1,6 +1,6 @@
http://www.w3.org/TR/html4/loose.dtd;>
-
+
@@ -19,45 +19,45 @@
}
//-->
var methods =
{"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":6,"i7":6,"i8":6,"i9":38,"i10":38,"i11":38,"i12":38,"i13":6,"i14":18,"i15":38,"i16":6,"i17":6,"i18":6};
-var tabs = {65535:["t0","All Methods"],2:["t2","Instance
Methods"],4:["t3","Abstract Methods"],16:["t5","Default
Methods"],32:["t6","Deprecated Methods"]};
+var tabs =
{65535:["t0","æææ¹æ³"],2:["t2","å®ä¾æ¹æ³"],4:["t3","æ½è±¡æ¹æ³"],16:["t5","é»è®¤æ¹æ³"],32:["t6","å·²è¿æ¶çæ¹æ³"]};
var altColor = "altColor";
var rowColor = "rowColor";
var tableTab = "tableTab";
var activeTableTab = "activeTableTab";
-JavaScript is disabled on your browser.
+æ¨çæµè§å¨å·²ç¦ç¨ JavaScriptã
-Skip navigation links
+è·³è¿å¯¼èªé¾æ¥
-
-Overview
-Package
-Class
-Use
-Tree
-Deprecated
-Index
-Help
+
+æ¦è§
+ç¨åºå
+ç±»
+使ç¨
+æ
+å·²è¿æ¶
+ç´¢å¼
+帮å©
-PrevClass
-NextClass
+ä¸ä¸ä¸ªç±»
+ä¸ä¸ä¸ªç±»
-Frames
-NoFrames
+æ¡æ¶
+æ æ¡æ¶
-AllClasses
+ææç±»
@@ -149,12 +149,12 @@ public interface
-
Method Summary
-
-All Methods Instance Methods Abstract Methods Default Methods Deprecated Methods
+æ¹æ³æ¦è¦
+
+æææ¹æ³ å®ä¾æ¹æ³ æ½è±¡æ¹æ³ é»è®¤æ¹æ³ å·²è¿æ¶çæ¹æ³
-Modifier and Type |
-Method and Description |
+éå®ç¬¦åç±»å |
+æ¹æ³å说æ |
byte[] |
@@ -202,7 +202,7 @@ public interface
long |
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bcb555af/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.WriterThread.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.WriterThread.html
b/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.WriterThread.html
index 568bd54..8a0b072 100644
---
a/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.WriterThread.html
+++
b/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.WriterThread.html
@@ -18,7 +18,7 @@
catch(err) {
}
//-->
-var methods = {"i0":10,"i1":10,"i2":10};
+var methods = {"i0":10,"i1":10,"i2":10,"i3":10};
var tabs = {65535:["t0","All Methods"],2:["t2","Instance
Methods"],8:["t4","Concrete Methods"]};
var altColor = "altColor";
var rowColor = "rowColor";
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
-class BucketCache.WriterThread
+class BucketCache.WriterThread
extends HasThread
@@ -193,6 +193,13 @@ extends
+private void
+putIntoBackingMap(BlockCacheKeykey,
+ BucketCache.BucketEntrybucketEntry)
+Put the new bucket entry into backingMap.
+
+
+
void
run()
@@ -231,7 +238,7 @@ extends
inputQueue
-private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/BlockingQueue.html?is-external=true;
title="class or interface in java.util.concurrent">BlockingQueueBucketCache.RAMQueueEntry inputQueue
+private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/BlockingQueue.html?is-external=true;
title="class or interface in java.util.concurrent">BlockingQueueBucketCache.RAMQueueEntry inputQueue
@@ -240,7 +247,7 @@ extends
writerEnabled
-private volatileboolean writerEnabled
+private volatileboolean writerEnabled
@@ -257,7 +264,7 @@ extends
WriterThread
-WriterThread(https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/BlockingQueue.html?is-external=true;
title="class or interface in java.util.concurrent">BlockingQueueBucketCache.RAMQueueEntryqueue)
+WriterThread(https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/BlockingQueue.html?is-external=true;
title="class or interface in java.util.concurrent">BlockingQueueBucketCache.RAMQueueEntryqueue)
@@ -274,7 +281,7 @@ extends
disableWriter
-voiddisableWriter()
+voiddisableWriter()
@@ -283,7 +290,7 @@ extends
run
-publicvoidrun()
+publicvoidrun()
Specified by:
https://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true#run--;
title="class or interface in java.lang">runin
interfacehttps://docs.oracle.com/javase/8/docs/api/java/lang/Runnable.html?is-external=true;
title="class or interface in java.lang">Runnable
@@ -292,13 +299,37 @@ extends
+
+
+
+
+putIntoBackingMap
+privatevoidputIntoBackingMap(BlockCacheKeykey,
+ BucketCache.BucketEntrybucketEntry)
+Put the new bucket entry into backingMap. Notice that we
are allowed to replace the existing
+ cache with a new block for the same cache key. there's a corner case: one
thread cache a
+ block in ramCache, copy to io-engine and add a bucket entry to backingMap.
Caching another
+ new block with the same cache key do the same thing for the same cache key,
so if not evict
+ the previous bucket entry, then memory leak happen because the previous
bucketEntry is gone
+ but the bucketAllocator do not free its memory.
+
+Parameters:
+key - Block cache key
+bucketEntry - Bucket entry to put into backingMap.
+See Also:
+BlockCacheUtil.shouldReplaceExistingCacheBlock(BlockCache
blockCache,BlockCacheKey
+ cacheKey, Cacheable newBlock)
+
+
+
doDrain
-voiddoDrain(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListBucketCache.RAMQueueEntryentries)
+voiddoDrain(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListBucketCache.RAMQueueEntryentries)
throws https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true;
title="class or interface in java.lang">InterruptedException
Flush the entries in ramCache to IOEngine and add bucket
entry to backingMap.
Process all that are passed in even if failure being sure to remove from
ramCache else we'll
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bcb555af/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.html
b/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.html
index 11df7d7..dccbeab 100644
--- a/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.html
+++ b/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.html
@@ -18,7 +18,7 @@
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a5c66de0/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
b/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
index a2ca0d5..f7aa217 100644
--- a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
@@ -167,9 +167,9 @@
java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
title="class or interface in java.io">Serializable)
-org.apache.hadoop.hbase.backup.BackupInfo.BackupState
-org.apache.hadoop.hbase.backup.BackupType
org.apache.hadoop.hbase.backup.BackupRestoreConstants.BackupCommand
+org.apache.hadoop.hbase.backup.BackupType
+org.apache.hadoop.hbase.backup.BackupInfo.BackupState
org.apache.hadoop.hbase.backup.BackupInfo.BackupPhase
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a5c66de0/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
b/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
index a2848a1..510c7b7 100644
--- a/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
+++ b/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
@@ -3042,6 +3042,12 @@ service.
Called before the client performs a Get
+
+void
+MetaTableMetrics.ExampleRegionObserverMeta.preGetOp(ObserverContextRegionCoprocessorEnvironmente,
+Getget,
+https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListCellresults)
+
@@ -8152,8 +8158,12 @@ service.
private static Pairhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String,TablePermission
-AccessControlLists.parsePermissionRecord(byte[]entryName,
- Cellkv)
+AccessControlLists.parsePermissionRecord(byte[]entryName,
+ Cellkv,
+ byte[]cf,
+ byte[]cq,
+ booleanfilterPerms,
+ https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in
java.lang">StringfilterUser)
Cell
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a5c66de0/devapidocs/org/apache/hadoop/hbase/class-use/Coprocessor.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/Coprocessor.html
b/devapidocs/org/apache/hadoop/hbase/class-use/Coprocessor.html
index e3ca75f..198a6ee 100644
--- a/devapidocs/org/apache/hadoop/hbase/class-use/Coprocessor.html
+++ b/devapidocs/org/apache/hadoop/hbase/class-use/Coprocessor.html
@@ -317,6 +317,16 @@
class
+MetaTableMetrics
+A coprocessor that collects metrics from meta table.
+
+
+
+(package private) class
+MetaTableMetrics.ExampleRegionObserverMeta
+
+
+class
MultiRowMutationEndpoint
This class demonstrates how to implement atomic multi row
transactions using
HRegion.mutateRowsWithLocks(Collection,
Collection, long, long)
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a5c66de0/devapidocs/org/apache/hadoop/hbase/class-use/CoprocessorEnvironment.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/class-use/CoprocessorEnvironment.html
b/devapidocs/org/apache/hadoop/hbase/class-use/CoprocessorEnvironment.html
index d041b83..f970173 100644
--- a/devapidocs/org/apache/hadoop/hbase/class-use/CoprocessorEnvironment.html
+++ b/devapidocs/org/apache/hadoop/hbase/class-use/CoprocessorEnvironment.html
@@ -323,13 +323,17 @@
void
+MetaTableMetrics.start(CoprocessorEnvironmentenv)
+
+
+void
BaseRowProcessorEndpoint.start(CoprocessorEnvironmentenv)
Stores a reference to the coprocessor environment provided
by the
RegionCoprocessorHost
from the region where this
coprocessor is loaded.
-
+
void
AggregateImplementation.start(CoprocessorEnvironmentenv)
Stores a reference to the coprocessor environment provided
by the
@@ -337,14 +341,18 @@
coprocessor is loaded.
-
+
void
Export.start(CoprocessorEnvironmentenvironment)
-
+
void
MultiRowMutationEndpoint.stop(CoprocessorEnvironmentenv)
+
+void
+MetaTableMetrics.stop(CoprocessorEnvironmente)
+
void
BaseRowProcessorEndpoint.stop(CoprocessorEnvironmentenv)
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb5d2c62/book.html
--
diff --git a/book.html b/book.html
index 33f6b3f..9483e80 100644
--- a/book.html
+++ b/book.html
@@ -162,173 +162,180 @@
91. Limitations of the Backup and Restore
Utility
+Synchronous Replication
+
+92. Background
+93. Design
+94. Operation and
maintenance
+
+
Apache HBase APIs
-92. Examples
+95. Examples
Apache HBase External APIs
-93. REST
-94. Thrift
-95. C/C++ Apache HBase Client
-96. Using Java Data Objects (JDO) with HBase
-97. Scala
-98. Jython
+96. REST
+97. Thrift
+98. C/C++ Apache HBase Client
+99. Using Java Data Objects (JDO) with HBase
+100. Scala
+101. Jython
Thrift API and Filter Language
-99. Filter Language
+102. Filter Language
HBase and Spark
-100. Basic Spark
-101. Spark Streaming
-102. Bulk Load
-103. SparkSQL/DataFrames
+103. Basic Spark
+104. Spark Streaming
+105. Bulk Load
+106. SparkSQL/DataFrames
Apache HBase Coprocessors
-104. Coprocessor Overview
-105. Types of Coprocessors
-106. Loading Coprocessors
-107. Examples
-108. Guidelines For
Deploying A Coprocessor
-109. Restricting Coprocessor
Usage
+107. Coprocessor Overview
+108. Types of Coprocessors
+109. Loading Coprocessors
+110. Examples
+111. Guidelines For
Deploying A Coprocessor
+112. Restricting Coprocessor
Usage
Apache HBase Performance Tuning
-110. Operating System
-111. Network
-112. Java
-113. HBase Configurations
-114. ZooKeeper
-115. Schema Design
-116. HBase General Patterns
-117. Writing to HBase
-118. Reading from HBase
-119. Deleting from HBase
-120. HDFS
-121. Amazon EC2
-122. Collocating HBase and
MapReduce
-123. Case Studies
+113. Operating System
+114. Network
+115. Java
+116. HBase Configurations
+117. ZooKeeper
+118. Schema Design
+119. HBase General Patterns
+120. Writing to HBase
+121. Reading from HBase
+122. Deleting from HBase
+123. HDFS
+124. Amazon EC2
+125. Collocating HBase and
MapReduce
+126. Case Studies
Troubleshooting and Debugging Apache HBase
-124. General Guidelines
-125. Logs
-126. Resources
-127. Tools
-128. Client
-129. MapReduce
-130. NameNode
-131. Network
-132. RegionServer
-133. Master
-134. ZooKeeper
-135. Amazon EC2
-136. HBase and Hadoop version issues
-137. HBase and HDFS
-138. Running unit or integration tests
-139. Case Studies
-140. Cryptographic Features
-141. Operating System
Specific Issues
-142. JDK Issues
+127. General Guidelines
+128. Logs
+129. Resources
+130. Tools
+131. Client
+132. MapReduce
+133. NameNode
+134. Network
+135. RegionServer
+136. Master
+137. ZooKeeper
+138. Amazon EC2
+139. HBase and Hadoop version issues
+140. HBase and HDFS
+141. Running unit or integration tests
+142. Case Studies
+143. Cryptographic Features
+144. Operating System
Specific Issues
+145. JDK Issues
Apache HBase Case Studies
-143. Overview
-144. Schema Design
-145. Performance/Troubleshooting
+146. Overview
+147. Schema Design
+148. Performance/Troubleshooting
Apache HBase Operational Management
-146. HBase Tools and Utilities
-147. Region Management
-148. Node Management
-149. HBase Metrics
-150. HBase Monitoring
-151. Cluster Replication
-152. Running
Multiple Workloads On a Single Cluster
-153. HBase Backup
-154. HBase Snapshots
-155. Storing Snapshots in Microsoft Azure Blob
Storage
-156. Capacity Planning and Region Sizing
-157. Table Rename
-158. RegionServer Grouping
-159. Region Normalizer
+149. HBase Tools and Utilities
+150. Region Management
+151. Node Management
+152. HBase Metrics
+153. HBase Monitoring
+154. Cluster Replication
+155. Running
Multiple Workloads On a Single Cluster
+156. HBase Backup
+157. HBase Snapshots
+158. Storing Snapshots in Microsoft Azure Blob
Storage
+159. Capacity Planning and Region Sizing
+160. Table Rename
+161. RegionServer Grouping
+162. Region Normalizer
Building and Developing Apache HBase
-160. Getting Involved
-161. Apache HBase Repositories
-162. IDEs
-163. Building Apache HBase
-164. Releasing Apache HBase
-165. Voting on Release Candidates
-166. Announcing Releases
-167. Generating the HBase Reference Guide
-168. Updating https://hbase.apache.org;>hbase.apache.org
-169. Tests
-170. Developer Guidelines
+163. Getting Involved
+164. Apache HBase Repositories
+165. IDEs
+166. Building Apache HBase
+167. Releasing Apache HBase
+168. Voting on Release Candidates
+169. Announcing Releases
+170. Generating the HBase Reference Guide
+171. Updating https://hbase.apache.org;>hbase.apache.org
+172. Tests
+173. Developer Guidelines
Unit Testing HBase Applications
-171. JUnit
-172. Mockito
-173. MRUnit
-174.
Integration Testing with an HBase Mini-Cluster
+174. JUnit
+175. Mockito
+176. MRUnit
+177.
Integration Testing with an HBase Mini-Cluster
Protobuf in HBase
-175. Protobuf
+178. Protobuf
Procedure Framework (Pv2): https://issues.apache.org/jira/browse/HBASE-12439;>HBASE-12439
-176. Procedures
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/55ce8d97/devapidocs/index-all.html
--
diff --git a/devapidocs/index-all.html b/devapidocs/index-all.html
index e8bb053..b317346 100644
--- a/devapidocs/index-all.html
+++ b/devapidocs/index-all.html
@@ -2123,7 +2123,7 @@
Add to online regions.
-addRegion(byte[],
int, int, int, int, long, long, long, int, int, int, long, long) -
Method in class org.apache.hadoop.hbase.rest.model.StorageClusterStatusModel.Node
+addRegion(byte[],
int, int, int, int, long, long, long, long, int, int, int, long,
long) - Method in class org.apache.hadoop.hbase.rest.model.StorageClusterStatusModel.Node
Add a region name to the list
@@ -10839,6 +10839,10 @@
checkRemoveBackupImages(FileSystem,
String, String[]) - Static method in class
org.apache.hadoop.hbase.backup.impl.BackupCommands.RepairCommand
+checkReopened(HRegionLocation)
- Method in class org.apache.hadoop.hbase.master.assignment.RegionStates
+
+Check whether the region has been reopened.
+
checkReplicaId(int)
- Static method in class org.apache.hadoop.hbase.client.RegionInfoBuilder.MutableRegionInfo
checkReplicationScope(ColumnFamilyDescriptor)
- Method in class org.apache.hadoop.hbase.master.HMaster
@@ -18544,10 +18548,40 @@
It is used internally. As
of release 2.0.0, this will be removed in HBase 3.0.0.
+CP_REQUEST_COST_KEY
- Static variable in class org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer.CPRequestCostFunction
+
+CP_REQUEST_COUNT
- Static variable in interface org.apache.hadoop.hbase.regionserver.MetricsRegionServerSource
+
+CP_REQUEST_COUNT
- Static variable in interface org.apache.hadoop.hbase.regionserver.MetricsTableSource
+
+CP_REQUEST_COUNT_DESC
- Static variable in interface org.apache.hadoop.hbase.regionserver.MetricsRegionServerSource
+
+CP_REQUEST_COUNT_DESC
- Static variable in interface org.apache.hadoop.hbase.regionserver.MetricsTableSource
+
cpEnv
- Variable in class org.apache.hadoop.hbase.quotas.MasterSpaceQuotaObserver
cpHost -
Variable in class org.apache.hadoop.hbase.master.HMaster
+CPRequestCostFunction(Configuration)
- Constructor for class org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer.CPRequestCostFunction
+
+cpRequestCount
- Variable in class org.apache.hadoop.hbase.RegionMetricsBuilder
+
+cpRequestCount
- Variable in class org.apache.hadoop.hbase.RegionMetricsBuilder.RegionMetricsImpl
+
+cpRequestsCount
- Variable in class org.apache.hadoop.hbase.master.balancer.BalancerRegionLoad
+
+cpRequestsCount
- Variable in class org.apache.hadoop.hbase.regionserver.HRegion
+
+cpRequestsCount
- Variable in class org.apache.hadoop.hbase.regionserver.MetricsRegionServerWrapperImpl
+
+cpRequestsCount
- Variable in class org.apache.hadoop.hbase.regionserver.MetricsTableWrapperAggregateImpl.MetricsTableValues
+
+cpRequestsCount
- Variable in class org.apache.hadoop.hbase.rest.model.StorageClusterStatusModel.Node.Region
+
+cpRequestsCount
- Variable in class org.apache.hadoop.hbase.ServerLoad
+
+Deprecated.
+
create(Connection,
String, BackupRequest) - Static method in class
org.apache.hadoop.hbase.backup.BackupClientFactory
create(Configuration)
- Static method in class org.apache.hadoop.hbase.backup.example.TableHFileArchiveTracker
@@ -20319,6 +20353,8 @@
createRegionFinder()
- Method in class org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer
+createRegionForReopen(RegionStates.RegionStateNode)
- Method in class org.apache.hadoop.hbase.master.assignment.RegionStates
+
createRegionInfos(TableDescriptor,
byte[][]) - Static method in class org.apache.hadoop.hbase.util.ModifyRegionUtils
createRegionLoad(HRegion,
ClusterStatusProtos.RegionLoad.Builder,
HBaseProtos.RegionSpecifier.Builder) - Method in class
org.apache.hadoop.hbase.regionserver.HRegionServer
@@ -20435,7 +20471,7 @@
createRemoteException(RPCProtos.ExceptionResponse)
- Static method in class org.apache.hadoop.hbase.ipc.IPCUtil
-createReopenProcedures(CollectionRegionInfo)
- Method in class org.apache.hadoop.hbase.master.assignment.AssignmentManager
+createReopenProcedure(MasterProcedureEnv,
HRegionLocation) - Method in class
org.apache.hadoop.hbase.master.procedure.ReopenTableRegionsProcedure
createReorderingProxy(ClientProtocol,
HFileSystem.ReorderBlocks, Configuration) - Static method in class
org.apache.hadoop.hbase.fs.HFileSystem
@@ -22753,6 +22789,8 @@
DEFAULT_COPY_MANIFEST_THREADS
- Static variable in class org.apache.hadoop.hbase.snapshot.ExportSnapshot
+DEFAULT_CP_REQUEST_COST
- Static variable in class org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer.CPRequestCostFunction
+
DEFAULT_CREATE_ZNODE_ATTEMPT_MAX_SLEEP_TIME
- Static variable in class org.apache.hadoop.hbase.util.HBaseFsck
DEFAULT_CREATE_ZNODE_ATTEMPT_SLEEP_INTERVAL
- Static variable in class org.apache.hadoop.hbase.util.HBaseFsck
@@ -23064,6
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2b11656f/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html
b/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html
index 75a99c7..79ec538 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html
@@ -151,7 +151,7 @@ extends
Nested classes/interfaces inherited from
classorg.apache.hadoop.hbase.master.HMaster
-HMaster.MasterStoppedException, HMaster.RedirectServlet
+HMaster.MasterStoppedException, HMaster.RedirectServlet,
HMaster.TableDescriptorGetter
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2b11656f/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.html
b/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.html
index 38f04e5..7a5ac2b 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.html
@@ -49,7 +49,7 @@ var activeTableTab = "activeTableTab";
-PrevClass
+PrevClass
NextClass
@@ -430,7 +430,7 @@ extends
-PrevClass
+PrevClass
NextClass
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2b11656f/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
b/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
index 8875aef..d027c3e 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/MasterRpcServices.BalanceSwitchMode.html
@@ -236,7 +236,7 @@ the order they are declared.
values
-public staticMasterRpcServices.BalanceSwitchMode[]values()
+public staticMasterRpcServices.BalanceSwitchMode[]values()
Returns an array containing the constants of this enum
type, in
the order they are declared. This method may be used to iterate
over the constants as follows:
@@ -256,7 +256,7 @@ for (MasterRpcServices.BalanceSwitchMode c :
MasterRpcServices.BalanceSwitchMode
valueOf
-public staticMasterRpcServices.BalanceSwitchModevalueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">Stringname)
+public staticMasterRpcServices.BalanceSwitchModevalueOf(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">Stringname)
Returns the enum constant of this type with the specified
name.
The string must match exactly an identifier used to declare an
enum constant in this type. (Extraneous whitespace characters are
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2b11656f/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.AssignmentProcedureEvent.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.AssignmentProcedureEvent.html
b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.AssignmentProcedureEvent.html
index dab6c6f..003530c 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.AssignmentProcedureEvent.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.AssignmentProcedureEvent.html
@@ -112,7 +112,7 @@
-private static class RegionStates.AssignmentProcedureEvent
+private static class RegionStates.AssignmentProcedureEvent
extends ProcedureEventRegionInfo
@@ -177,7 +177,7 @@ extends
AssignmentProcedureEvent
-publicAssignmentProcedureEvent(RegionInforegionInfo)
+publicAssignmentProcedureEvent(RegionInforegionInfo)
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2b11656f/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html
b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html
index 23f08b4..4bc5cd4 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/assignment/RegionStates.RegionFailedOpen.html
@@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab";
-public static final class
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/65565d77/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
index c6822d1..2405d9a 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/HMaster.html
@@ -18,7 +18,7 @@
catch(err) {
}
//-->
-var methods =
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":9,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":9,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":9,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":9,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109"
:10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":9,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":9,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":10,"i145":10,"i146":10,"i147":10,"i148":10,"i149":10,"i150":10,"i151":10,"i152":10,"i153":10,"i154":10,"i155":10,"i156":10,"i157":10,"i158":10,"i159":10,"i160":10,"i161":10,"i162":10,"i163":10,"i164":10,"i165":10,"i166":10,"i167":10,"i168":10,"i169":10,"i170":10,"i171":10,"i172":9};
+var methods =
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":9,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":9,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":9,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":9,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109"
:10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":9,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":9,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":10,"i145":10,"i146":10,"i147":10,"i148":10,"i149":10,"i150":10,"i151":10,"i152":10,"i153":10,"i154":10,"i155":10,"i156":10,"i157":10,"i158":10,"i159":10,"i160":10,"i161":10,"i162":10,"i163":10,"i164":10,"i165":10,"i166":10,"i167":10,"i168":10,"i169":9};
var tabs = {65535:["t0","All Methods"],1:["t1","Static
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
var altColor = "altColor";
var rowColor = "rowColor";
@@ -128,7 +128,7 @@ var activeTableTab = "activeTableTab";
@InterfaceAudience.LimitedPrivate(value="Tools")
-public class HMaster
+public class HMaster
extends HRegionServer
implements MasterServices
HMaster is the "master server" for HBase. An HBase cluster
has one active
@@ -422,42 +422,38 @@ implements rsFatals
-private ProcedureEvent?
-serverCrashProcessingEnabled
-
-
private ServerManager
serverManager
-
+
(package private) boolean
serviceStarted
-
+
(package private) SnapshotManager
snapshotManager
-
+
private SnapshotQuotaObserverChore
snapshotQuotaChore
-
+
private SpaceQuotaSnapshotNotifier
spaceQuotaSnapshotNotifier
-
+
private SplitOrMergeTracker
splitOrMergeTracker
-
+
private long
splitPlanCount
-
+
private TableStateManager
tableStateManager
-
+
private MasterWalManager
walManager
@@ -649,27 +645,31 @@ implements
+private void
+createProcedureExecutor()
+
+
private SpaceQuotaSnapshotNotifier
createQuotaSnapshotNotifier()
-
+
protected RSRpcServices
createRpcServices()
-
+
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e11cf2cb/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html
b/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html
index eb945fb..ad40e69 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/balancer/package-tree.html
@@ -197,8 +197,8 @@
java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
title="class or interface in java.io">Serializable)
-org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action.Type
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.LocalityType
+org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.Action.Type
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e11cf2cb/devapidocs/org/apache/hadoop/hbase/master/class-use/MasterServices.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/class-use/MasterServices.html
b/devapidocs/org/apache/hadoop/hbase/master/class-use/MasterServices.html
index 76ca590..559c744 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/class-use/MasterServices.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/class-use/MasterServices.html
@@ -382,9 +382,10 @@
TableNamespaceManager.createNamespaceTable(MasterServicesmasterServices)
-(package private) ServerManager
+protected ServerManager
HMaster.createServerManager(MasterServicesmaster)
-Create a ServerManager
instance.
+
+ Create a ServerManager
instance.
@@ -450,19 +451,15 @@
-ServerManager(MasterServicesmaster,
- booleanconnect)
-
-
SplitLogManager(MasterServicesmaster,
org.apache.hadoop.conf.Configurationconf)
Its OK to construct this object even when region-servers
are not online.
-
+
TableNamespaceManager(MasterServicesmasterServices)
-
+
TableStateManager(MasterServicesmaster)
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e11cf2cb/devapidocs/org/apache/hadoop/hbase/master/class-use/RegionServerTracker.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/class-use/RegionServerTracker.html
b/devapidocs/org/apache/hadoop/hbase/master/class-use/RegionServerTracker.html
index 2a8bf6c..e51f116 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/class-use/RegionServerTracker.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/class-use/RegionServerTracker.html
@@ -103,7 +103,7 @@
-(package private) RegionServerTracker
+private RegionServerTracker
HMaster.regionServerTracker
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e11cf2cb/devapidocs/org/apache/hadoop/hbase/master/class-use/ServerManager.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/class-use/ServerManager.html
b/devapidocs/org/apache/hadoop/hbase/master/class-use/ServerManager.html
index 7d1b3ed..db31e5d 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/class-use/ServerManager.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/class-use/ServerManager.html
@@ -128,9 +128,10 @@
-(package private) ServerManager
+protected ServerManager
HMaster.createServerManager(MasterServicesmaster)
-Create a ServerManager
instance.
+
+ Create a ServerManager
instance.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e11cf2cb/devapidocs/org/apache/hadoop/hbase/master/package-summary.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/master/package-summary.html
b/devapidocs/org/apache/hadoop/hbase/master/package-summary.html
index 81462ad..46e06a9 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/package-summary.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/package-summary.html
@@ -477,7 +477,8 @@
RegionServerTracker
-Tracks the online region servers via ZK.
+
+ Tracks the online region servers via ZK.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e11cf2cb/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
index 23075b7..e37f9f8 100644
---
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/3469cbc0/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
b/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
index 2e64b09..fa29c2b 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/assignment/AssignmentManager.html
@@ -18,7 +18,7 @@
catch(err) {
}
//-->
-var methods =
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i
109":10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10};
+var methods =
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i
109":10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10};
var tabs = {65535:["t0","All Methods"],2:["t2","Instance
Methods"],8:["t4","Concrete Methods"]};
var altColor = "altColor";
var rowColor = "rowColor";
@@ -611,69 +611,75 @@ implements RegionInforegionInfo)
+(package private) boolean
+isDeadServerProcessed(ServerNameserverName)
+This is a very particular check.
+
+
+
boolean
isFailoverCleanupDone()
Used by ServerCrashProcedure to make sure AssignmentManager
has completed
the failover cleanup before re-assigning regions of dead servers.
-
+
boolean
isMetaInitialized()
-
+
boolean
isMetaLoaded()
-
+
boolean
isMetaRegion(byte[]regionName)
-
+
private boolean
isMetaRegion(RegionInforegionInfo)
-
+
boolean
isMetaRegionInTransition()
-
+
boolean
isRunning()
-
+
boolean
isTableDisabled(TableNametableName)
-
+
boolean
isTableEnabled(TableNametableName)
-
+
void
joinCluster()
-
-void
+
+private void
killRegionServer(RegionStates.ServerStateNodeserverNode)
-
-void
+
+private void
killRegionServer(ServerNameserverName)
-
+
private void
loadMeta()
-
+
void
markRegionAsClosed(RegionStates.RegionStateNoderegionNode)
-
+
void
markRegionAsClosing(RegionStates.RegionStateNoderegionNode)
-
+
void
markRegionAsMerged(RegionInfochild,
ServerNameserverName,
@@ -682,73 +688,73 @@ implements When called here, the merge has happened.
-
+
void
markRegionAsOpened(RegionStates.RegionStateNoderegionNode)
-
+
void
markRegionAsOpening(RegionStates.RegionStateNoderegionNode)
-
+
void
markRegionAsSplit(RegionInfoparent,
ServerNameserverName,
RegionInfodaughterA,
RegionInfodaughterB)
-
+
void
move(RegionInforegionInfo)
-
+
https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/Future.html?is-external=true;
title="class or interface in
java.util.concurrent">Futurebyte[]
moveAsync(RegionPlanregionPlan)
-
+
void
offlineRegion(RegionInforegionInfo)
-
+
void
onlineRegion(RegionInforegionInfo,
ServerNameserverName)
-
+
private void
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4df09ed9/devapidocs/org/apache/hadoop/hbase/master/procedure/ReopenTableRegionsProcedure.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/procedure/ReopenTableRegionsProcedure.html
b/devapidocs/org/apache/hadoop/hbase/master/procedure/ReopenTableRegionsProcedure.html
new file mode 100644
index 000..c869777
--- /dev/null
+++
b/devapidocs/org/apache/hadoop/hbase/master/procedure/ReopenTableRegionsProcedure.html
@@ -0,0 +1,640 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+
+ReopenTableRegionsProcedure (Apache HBase 3.0.0-SNAPSHOT API)
+
+
+
+
+
+var methods =
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10};
+var tabs = {65535:["t0","All Methods"],2:["t2","Instance
Methods"],8:["t4","Concrete Methods"]};
+var altColor = "altColor";
+var rowColor = "rowColor";
+var tableTab = "tableTab";
+var activeTableTab = "activeTableTab";
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+PrevClass
+NextClass
+
+
+Frames
+NoFrames
+
+
+AllClasses
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.hadoop.hbase.master.procedure
+Class
ReopenTableRegionsProcedure
+
+
+
+https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">java.lang.Object
+
+
+org.apache.hadoop.hbase.procedure2.ProcedureTEnvironment
+
+
+org.apache.hadoop.hbase.procedure2.StateMachineProcedureMasterProcedureEnv,TState
+
+
+org.apache.hadoop.hbase.master.procedure.AbstractStateMachineTableProcedureorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ReopenTableRegionsState
+
+
+org.apache.hadoop.hbase.master.procedure.ReopenTableRegionsProcedure
+
+
+
+
+
+
+
+
+
+
+
+
+
+All Implemented Interfaces:
+https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableProcedureMasterProcedureEnv, TableProcedureInterface
+
+
+
+@InterfaceAudience.Private
+public class ReopenTableRegionsProcedure
+extends AbstractStateMachineTableProcedureorg.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ReopenTableRegionsState
+Used for non table procedures to reopen the regions for a
table. For example,
+ ModifyPeerProcedure.
+
+
+
+
+
+
+
+
+
+
+
+Nested Class Summary
+
+
+
+
+Nested classes/interfaces inherited from
classorg.apache.hadoop.hbase.procedure2.StateMachineProcedure
+StateMachineProcedure.Flow
+
+
+
+
+
+Nested classes/interfaces inherited from
classorg.apache.hadoop.hbase.procedure2.Procedure
+Procedure.LockState
+
+
+
+
+
+Nested classes/interfaces inherited from
interfaceorg.apache.hadoop.hbase.master.procedure.TableProcedureInterface
+TableProcedureInterface.TableOperationType
+
+
+
+
+
+
+
+
+Field Summary
+
+Fields
+
+Modifier and Type
+Field and Description
+
+
+private static org.slf4j.Logger
+LOG
+
+
+private TableName
+tableName
+
+
+
+
+
+
+Fields inherited from classorg.apache.hadoop.hbase.procedure2.Procedure
+NO_PROC_ID,
NO_TIMEOUT
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Constructor and Description
+
+
+ReopenTableRegionsProcedure()
+
+
+ReopenTableRegionsProcedure(TableNametableName)
+
+
+
+
+
+
+
+
+
+Method Summary
+
+All MethodsInstance MethodsConcrete Methods
+
+Modifier and Type
+Method and Description
+
+
+protected void
+deserializeStateData(ProcedureStateSerializerserializer)
+Called on store load to allow the user to decode the
previously serialized
+ state.
+
+
+
+protected StateMachineProcedure.Flow
+executeFromState(MasterProcedureEnvenv,
+
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ReopenTableRegionsStatestate)
+called to perform a single step of the specified 'state' of
the procedure
+
+
+
+protected
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ReopenTableRegionsState
+getInitialState()
+Return the initial state object that will be used for the
first call to executeFromState().
+
+
+
+protected
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.ReopenTableRegionsState
+getState(intstateId)
+Convert an ordinal (or state id) to an Enum (or more
descriptive) state object.
+
+
+
+protected int
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/883dde2f/devapidocs/org/apache/hadoop/hbase/package-summary.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/package-summary.html
b/devapidocs/org/apache/hadoop/hbase/package-summary.html
index afe234c..84822ae 100644
--- a/devapidocs/org/apache/hadoop/hbase/package-summary.html
+++ b/devapidocs/org/apache/hadoop/hbase/package-summary.html
@@ -291,12 +291,19 @@
+BBKVComparator
+
+A comparator for case where ByteBufferKeyValue is prevalent
type (BBKV
+ is base-type in hbase2).
+
+
+
ByteBufferExtendedCell
This class is a server side extension to the Cell interface.
-
+
ByteBufferKeyOnlyKeyValue
This is a key only Cell implementation which is identical
to KeyValue.KeyOnlyKeyValue
@@ -304,223 +311,223 @@
(onheap and offheap).
-
+
ByteBufferKeyValue
This Cell is an implementation of ByteBufferExtendedCell where the data
resides in
off heap/ on heap ByteBuffer
-
+
ByteBufferTag
This is a Tag implementation in which value is
backed by
https://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
title="class or interface in java.nio">ByteBuffer
-
+
CacheEvictionStats
-
+
CacheEvictionStatsAggregator
Used to merge CacheEvictionStats.
-
+
CacheEvictionStatsBuilder
-
+
CellBuilderFactory
Create a CellBuilder instance.
-
+
CellComparatorImpl
Compare two HBase cells.
-
+
CellComparatorImpl.MetaCellComparator
A CellComparatorImpl for
hbase:meta catalog table
KeyValues.
-
+
CellUtil
Utility methods helpful for slinging Cell instances.
-
+
ChoreService
ChoreService is a service that can be used to schedule
instances of ScheduledChore to run
periodically while sharing threads.
-
+
ChoreService.ChoreServiceThreadFactory
Custom ThreadFactory used with the
ScheduledThreadPoolExecutor so that all the threads are
daemon threads, and thus, don't prevent the JVM from shutting down
-
+
ClusterId
The identifier for this cluster.
-
+
ClusterMetricsBuilder
-
+
ClusterMetricsBuilder.ClusterMetricsImpl
-
+
ClusterStatus
Deprecated
As of release 2.0.0, this
will be removed in HBase 3.0.0
Use ClusterMetrics
instead.
-
+
CompatibilityFactory
Class that will create many instances of classes provided
by the hbase-hadoop{1|2}-compat jars.
-
+
CompatibilitySingletonFactory
Factory for classes supplied by hadoop compatibility
modules.
-
+
CompoundConfiguration
Do a shallow merge of multiple KV configuration pools.
-
+
CompoundConfiguration.ImmutableConfWrapper
-
+
DaemonThreadFactory
Thread factory that creates daemon threads
-
+
ExtendedCellBuilderFactory
-
+
ExtendedCellBuilderImpl
-
+
HBaseConfiguration
Adds HBase configuration files to a Configuration
-
+
HBaseInterfaceAudience
This class defines constants for different classes of hbase
limited private apis
-
+
HColumnDescriptor
Deprecated
-
+
HConstants
HConstants holds a bunch of HBase-related constants
-
+
HDFSBlocksDistribution
Data structure to describe the distribution of HDFS blocks
among hosts.
-
+
HDFSBlocksDistribution.HostAndWeight
Stores the hostname and weight for that hostname.
-
+
HDFSBlocksDistribution.HostAndWeight.WeightComparator
comparator used to sort hosts based on weight
-
+
HealthCheckChore
The Class HealthCheckChore for running health checker
regularly.
-
+
HealthChecker
A utility for executing an external script that checks the
health of
the node.
-
+
HealthReport
The Class HealthReport containing information about health
of the node.
-
+
HRegionInfo
Deprecated
As of release 2.0.0, this
will be removed in HBase 3.0.0.
-
+
HRegionLocation
Data structure to hold RegionInfo and the address for the
hosting
HRegionServer.
-
+
HTableDescriptor
Deprecated
As of release 2.0.0, this
will be removed in HBase 3.0.0.
-
+
IndividualBytesFieldCell
-
+
IndividualBytesFieldCellBuilder
-
+
JitterScheduledThreadPoolExecutorImpl
ScheduledThreadPoolExecutor that will add some jitter to
the RunnableScheduledFuture.getDelay.
-
+
JMXListener
Pluggable JMX Agent for HBase(to fix the 2 random TCP ports
issue
@@ -530,336 +537,336 @@
3)support subset of SSL (with default configuration)
-
+
KeyValue
An HBase Key/Value.
-
+
KeyValue.KeyOnlyKeyValue
A simple form of KeyValue that creates a keyvalue with only
the key part of the byte[]
Mainly used in places where we need to compare two cells.
-
+
KeyValue.KVComparator
Deprecated
: Use CellComparatorImpl.
-
+
KeyValue.MetaComparator
Deprecated
: CellComparatorImpl.META_COMPARATOR
to be used.
-
+
KeyValueBuilder
-
+
KeyValueTestUtil
-
+
KeyValueUtil
static convenience methods for dealing with
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7bcc960d/devapidocs/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html
b/devapidocs/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html
index 51a2cee..08907ec 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html
@@ -316,8 +316,9 @@ extends
void
-postCompletedModifyTableAction(TableNametableName,
- TableDescriptorhtd,
+postCompletedModifyTableAction(TableNametableName,
+ TableDescriptoroldDescriptor,
+ TableDescriptorcurrentDescriptor,
Useruser)
@@ -441,12 +442,14 @@ extends
void
-postModifyNamespace(NamespaceDescriptorns)
+postModifyNamespace(NamespaceDescriptoroldNsDescriptor,
+ NamespaceDescriptorcurrentNsDescriptor)
void
-postModifyTable(TableNametableName,
- TableDescriptorhtd)
+postModifyTable(TableNametableName,
+ TableDescriptoroldDescriptor,
+ TableDescriptorcurrentDescriptor)
void
@@ -755,17 +758,20 @@ extends
void
-preModifyNamespace(NamespaceDescriptorns)
+preModifyNamespace(NamespaceDescriptorcurrentNsDescriptor,
+ NamespaceDescriptornewNsDescriptor)
void
-preModifyTable(TableNametableName,
- TableDescriptorhtd)
+preModifyTable(TableNametableName,
+ TableDescriptorcurrentDescriptor,
+ TableDescriptornewDescriptor)
void
-preModifyTableAction(TableNametableName,
-TableDescriptorhtd,
+preModifyTableAction(TableNametableName,
+TableDescriptorcurrentDescriptor,
+TableDescriptornewDescriptor,
Useruser)
@@ -1101,13 +1107,14 @@ extends
+
preModifyNamespace
-publicvoidpreModifyNamespace(NamespaceDescriptorns)
+publicvoidpreModifyNamespace(NamespaceDescriptorcurrentNsDescriptor,
+ NamespaceDescriptornewNsDescriptor)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
Throws:
@@ -1115,13 +1122,14 @@ extends
+
postModifyNamespace
-publicvoidpostModifyNamespace(NamespaceDescriptorns)
+publicvoidpostModifyNamespace(NamespaceDescriptoroldNsDescriptor,
+NamespaceDescriptorcurrentNsDescriptor)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
Throws:
@@ -1135,7 +1143,7 @@ extends
preGetNamespaceDescriptor
-publicvoidpreGetNamespaceDescriptor(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">StringnamespaceName)
+publicvoidpreGetNamespaceDescriptor(https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">StringnamespaceName)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
Throws:
@@ -1149,7 +1157,7 @@ extends
postGetNamespaceDescriptor
-publicvoidpostGetNamespaceDescriptor(NamespaceDescriptorns)
+publicvoidpostGetNamespaceDescriptor(NamespaceDescriptorns)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
Throws:
@@ -1163,7 +1171,7 @@ extends
preListNamespaceDescriptors
-publicvoidpreListNamespaceDescriptors(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListNamespaceDescriptordescriptors)
+publicvoidpreListNamespaceDescriptors(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListNamespaceDescriptordescriptors)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
Throws:
@@ -1177,7 +1185,7 @@ extends
postListNamespaceDescriptors
-publicvoidpostListNamespaceDescriptors(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListNamespaceDescriptordescriptors)
+publicvoidpostListNamespaceDescriptors(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListNamespaceDescriptordescriptors)
throws
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/92a26cfb/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html
index 7ec6aa5..e2dd09e 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html
@@ -18,7 +18,7 @@
catch(err) {
}
//-->
-var methods =
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":9,"i42":9,"i43":9,"i44":9,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":9,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":9,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":9,"i106":10,"i107":10,"i108":10,"i109":10
,"i110":10,"i111":10,"i112":41,"i113":41,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":9,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":10,"i145":10,"i146":42,"i147":10,"i148":10,"i149":10,"i150":10,"i151":10,"i152":10,"i153":10,"i154":10,"i155":10,"i156":10,"i157":10,"i158":10,"i159":10,"i160":10,"i161":10,"i162":10,"i163":10,"i164":10,"i165":10,"i166":10,"i167":10,"i168":9,"i169":10,"i170":10,"i171":10,"i172":10,"i173":10,"i174":10,"i175":10,"i176":10,"i177":9,"i178":10,"i179":10,"i180":9,"i181":9,"i182":9,"i183":9,"i184":9,"i185":9,"i186":9,"i187":9,"i188":9,"i189":9,"i190":10,"i191":10,"i192":10,"i193":10,"i194":10,"i195":10,"i196":10,"i197":10,"i198":10,"i199":9,"i200":10,"i201":10,"i202":10,"i203":10,"i204":10,"i205":10,"i206":10,"i207":10,"i208":10,"i209":10,"i210":10,
"i211":10,"i212":10,"i213":10,"i214":10,"i215":10,"i216":10,"i217":10,"i218":10,"i219":10,"i220":10,"i221":10,"i222":10,"i223":10,"i224":10,"i225":10,"i226":10,"i227":10,"i228":10,"i229":10,"i230":9,"i231":9,"i232":10,"i233":10,"i234":10,"i235":10,"i236":10,"i237":10,"i238":10,"i239":10,"i240":10,"i241":10,"i242":10,"i243":9,"i244":10,"i245":10,"i246":10,"i247":10,"i248":10,"i249":10,"i250":10,"i251":10,"i252":10,"i253":10,"i254":10,"i255":10,"i256":10,"i257":9,"i258":10,"i259":10,"i260":10,"i261":10};
+var methods =
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":9,"i42":9,"i43":9,"i44":9,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":9,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":9,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":9,"i107":10,"i108":10,"i109":10
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.ReplicationBarrierResult.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.ReplicationBarrierResult.html
b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.ReplicationBarrierResult.html
index cdd2f36..fea2b5a 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.ReplicationBarrierResult.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.ReplicationBarrierResult.html
@@ -151,2029 +151,2019 @@
143 private static final Logger LOG =
LoggerFactory.getLogger(MetaTableAccessor.class);
144 private static final Logger METALOG =
LoggerFactory.getLogger("org.apache.hadoop.hbase.META");
145
-146 private static final byte[]
META_REGION_PREFIX;
-147 static {
-148// Copy the prefix from
FIRST_META_REGIONINFO into META_REGION_PREFIX.
-149// FIRST_META_REGIONINFO ==
'hbase:meta,,1'. META_REGION_PREFIX == 'hbase:meta,'
-150int len =
RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName().length - 2;
-151META_REGION_PREFIX = new byte
[len];
-152
System.arraycopy(RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName(), 0,
-153 META_REGION_PREFIX, 0, len);
-154 }
-155
-156 @VisibleForTesting
-157 public static final byte[]
REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent");
-158
-159 private static final byte ESCAPE_BYTE =
(byte) 0xFF;
-160
-161 private static final byte
SEPARATED_BYTE = 0x00;
-162
-163 /**
-164 * Lists all of the table regions
currently in META.
-165 * Deprecated, keep there until some
test use this.
-166 * @param connection what we will use
-167 * @param tableName table to list
-168 * @return Map of all user-space
regions to servers
-169 * @deprecated use {@link
#getTableRegionsAndLocations}, region can have multiple locations
-170 */
-171 @Deprecated
-172 public static
NavigableMapRegionInfo, ServerName allTableRegions(
-173 Connection connection, final
TableName tableName) throws IOException {
-174final NavigableMapRegionInfo,
ServerName regions = new TreeMap();
-175Visitor visitor = new
TableVisitorBase(tableName) {
-176 @Override
-177 public boolean visitInternal(Result
result) throws IOException {
-178RegionLocations locations =
getRegionLocations(result);
-179if (locations == null) return
true;
-180for (HRegionLocation loc :
locations.getRegionLocations()) {
-181 if (loc != null) {
-182RegionInfo regionInfo =
loc.getRegionInfo();
-183regions.put(regionInfo,
loc.getServerName());
-184 }
-185}
-186return true;
-187 }
-188};
-189scanMetaForTableRegions(connection,
visitor, tableName);
-190return regions;
-191 }
-192
-193 @InterfaceAudience.Private
-194 public enum QueryType {
-195ALL(HConstants.TABLE_FAMILY,
HConstants.CATALOG_FAMILY),
-196REGION(HConstants.CATALOG_FAMILY),
-197TABLE(HConstants.TABLE_FAMILY),
-198
REPLICATION(HConstants.REPLICATION_BARRIER_FAMILY);
-199
-200private final byte[][] families;
-201
-202QueryType(byte[]... families) {
-203 this.families = families;
-204}
-205
-206byte[][] getFamilies() {
-207 return this.families;
-208}
-209 }
-210
-211 /** The delimiter for meta columns for
replicaIds gt; 0 */
-212 protected static final char
META_REPLICA_ID_DELIMITER = '_';
-213
-214 /** A regex for parsing server columns
from meta. See above javadoc for meta layout */
-215 private static final Pattern
SERVER_COLUMN_PATTERN
-216=
Pattern.compile("^server(_[0-9a-fA-F]{4})?$");
-217
-218
-219 // Reading operations //
-220
-221
-222 /**
-223 * Performs a full scan of
codehbase:meta/code for regions.
-224 * @param connection connection we're
using
-225 * @param visitor Visitor invoked
against each row in regions family.
+146 @VisibleForTesting
+147 public static final byte[]
REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent");
+148
+149 private static final byte ESCAPE_BYTE =
(byte) 0xFF;
+150
+151 private static final byte
SEPARATED_BYTE = 0x00;
+152
+153 /**
+154 * Lists all of the table regions
currently in META.
+155 * Deprecated, keep there until some
test use this.
+156 * @param connection what we will use
+157 * @param tableName table to list
+158 * @return Map of all user-space
regions to servers
+159 * @deprecated use {@link
#getTableRegionsAndLocations}, region can have multiple locations
+160 */
+161 @Deprecated
+162 public static
NavigableMapRegionInfo, ServerName allTableRegions(
+163 Connection connection, final
TableName tableName) throws IOException {
+164final NavigableMapRegionInfo,
ServerName regions = new TreeMap();
+165
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/de18d468/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
index 59c95fc..bf2f539 100644
--- a/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/util/package-tree.html
@@ -532,14 +532,14 @@
java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
title="class or interface in java.io">Serializable)
-org.apache.hadoop.hbase.util.PoolMap.PoolType
-org.apache.hadoop.hbase.util.Order
-org.apache.hadoop.hbase.util.ChecksumType
+org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE
org.apache.hadoop.hbase.util.PrettyPrinter.Unit
+org.apache.hadoop.hbase.util.ChecksumType
org.apache.hadoop.hbase.util.IdReadWriteLock.ReferenceType
-org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE
-org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.UnsafeComparer
(implements org.apache.hadoop.hbase.util.Bytes.ComparerT)
+org.apache.hadoop.hbase.util.Order
org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.PureJavaComparer
(implements org.apache.hadoop.hbase.util.Bytes.ComparerT)
+org.apache.hadoop.hbase.util.PoolMap.PoolType
+org.apache.hadoop.hbase.util.Bytes.LexicographicalComparerHolder.UnsafeComparer
(implements org.apache.hadoop.hbase.util.Bytes.ComparerT)
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/de18d468/devapidocs/org/apache/hadoop/hbase/wal/class-use/WAL.Entry.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/wal/class-use/WAL.Entry.html
b/devapidocs/org/apache/hadoop/hbase/wal/class-use/WAL.Entry.html
index f862989..296589b 100644
--- a/devapidocs/org/apache/hadoop/hbase/wal/class-use/WAL.Entry.html
+++ b/devapidocs/org/apache/hadoop/hbase/wal/class-use/WAL.Entry.html
@@ -370,10 +370,6 @@ Input/OutputFormats, a table indexing MapReduce job, and
utility methods.
private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListWAL.Entry
-HBaseInterClusterReplicationEndpoint.Replicator.entries
-
-
-private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListWAL.Entry
WALEntryBatch.walEntries
@@ -412,7 +408,18 @@ Input/OutputFormats, a table indexing MapReduce job, and
utility methods.
private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">Listhttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListWAL.Entry
-HBaseInterClusterReplicationEndpoint.createBatches(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListWAL.Entryentries)
+HBaseInterClusterReplicationEndpoint.createBatches(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListWAL.Entryentries)
+Divide the entries into multiple batches, so that we can
replicate each batch in a thread pool
+ concurrently.
+
+
+
+private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">Listhttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListWAL.Entry
+HBaseInterClusterReplicationEndpoint.createParallelBatches(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListWAL.Entryentries)
+
+
+private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">Listhttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListWAL.Entry
+HBaseInterClusterReplicationEndpoint.createSerialBatches(https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListWAL.Entryentries)
private https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">Listhttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListWAL.Entry
@@ -460,16 +467,20 @@
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d220bc5e/devapidocs/org/apache/hadoop/hbase/client/AsyncAdmin.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/AsyncAdmin.html
b/devapidocs/org/apache/hadoop/hbase/client/AsyncAdmin.html
index f481b56..8618b6c 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/AsyncAdmin.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/AsyncAdmin.html
@@ -18,7 +18,7 @@
catch(err) {
}
//-->
-var methods =
{"i0":6,"i1":6,"i2":18,"i3":6,"i4":6,"i5":6,"i6":18,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":18,"i16":18,"i17":6,"i18":6,"i19":6,"i20":6,"i21":6,"i22":6,"i23":6,"i24":6,"i25":6,"i26":6,"i27":6,"i28":6,"i29":6,"i30":6,"i31":6,"i32":6,"i33":6,"i34":6,"i35":6,"i36":6,"i37":6,"i38":6,"i39":6,"i40":6,"i41":6,"i42":6,"i43":6,"i44":6,"i45":6,"i46":6,"i47":6,"i48":18,"i49":6,"i50":6,"i51":18,"i52":6,"i53":6,"i54":6,"i55":6,"i56":6,"i57":6,"i58":18,"i59":18,"i60":18,"i61":6,"i62":6,"i63":6,"i64":6,"i65":6,"i66":6,"i67":6,"i68":18,"i69":6,"i70":6,"i71":6,"i72":6,"i73":6,"i74":6,"i75":6,"i76":6,"i77":6,"i78":6,"i79":6,"i80":6,"i81":6,"i82":6,"i83":6,"i84":18,"i85":6,"i86":6,"i87":6,"i88":6,"i89":6,"i90":6,"i91":6,"i92":18,"i93":6,"i94":6,"i95":6,"i96":18,"i97":6,"i98":6,"i99":6,"i100":6,"i101":6,"i102":18,"i103":18,"i104":6,"i105":6,"i106":6,"i107":6,"i108":6,"i109":6,"i110":6,"i111":6,"i112":6,"i113":6,"i114":6,"i115":6,"i116":6,"i117":6,"i118":6,"i119"
:6,"i120":6,"i121":6,"i122":6,"i123":6,"i124":6,"i125":6,"i126":6,"i127":6,"i128":6,"i129":6,"i130":18,"i131":18,"i132":6,"i133":6,"i134":6,"i135":6,"i136":6,"i137":6,"i138":6,"i139":6,"i140":6,"i141":6,"i142":6,"i143":6,"i144":6};
+var methods =
{"i0":6,"i1":6,"i2":18,"i3":6,"i4":6,"i5":6,"i6":18,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":18,"i17":18,"i18":6,"i19":6,"i20":6,"i21":6,"i22":6,"i23":6,"i24":6,"i25":6,"i26":6,"i27":6,"i28":6,"i29":6,"i30":6,"i31":6,"i32":6,"i33":6,"i34":6,"i35":6,"i36":6,"i37":6,"i38":6,"i39":6,"i40":6,"i41":6,"i42":6,"i43":6,"i44":6,"i45":6,"i46":6,"i47":6,"i48":6,"i49":18,"i50":6,"i51":6,"i52":18,"i53":6,"i54":6,"i55":6,"i56":6,"i57":6,"i58":6,"i59":18,"i60":18,"i61":18,"i62":6,"i63":6,"i64":6,"i65":6,"i66":6,"i67":6,"i68":6,"i69":18,"i70":6,"i71":6,"i72":6,"i73":6,"i74":6,"i75":6,"i76":6,"i77":6,"i78":6,"i79":6,"i80":6,"i81":6,"i82":6,"i83":6,"i84":6,"i85":18,"i86":6,"i87":6,"i88":6,"i89":6,"i90":6,"i91":6,"i92":6,"i93":18,"i94":6,"i95":6,"i96":6,"i97":18,"i98":6,"i99":6,"i100":6,"i101":6,"i102":6,"i103":18,"i104":18,"i105":6,"i106":6,"i107":6,"i108":6,"i109":6,"i110":6,"i111":6,"i112":6,"i113":6,"i114":6,"i115":6,"i116":6,"i117":6,"i118":6,"i119"
:6,"i120":6,"i121":6,"i122":6,"i123":6,"i124":6,"i125":6,"i126":6,"i127":6,"i128":6,"i129":6,"i130":6,"i131":18,"i132":18,"i133":6,"i134":6,"i135":6,"i136":6,"i137":6,"i138":6,"i139":6,"i140":6,"i141":6,"i142":6,"i143":6,"i144":6,"i145":6};
var tabs = {65535:["t0","All Methods"],2:["t2","Instance
Methods"],4:["t3","Abstract Methods"],16:["t5","Default Methods"]};
var altColor = "altColor";
var rowColor = "rowColor";
@@ -227,19 +227,27 @@ public interface
+https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
title="class or interface in java.lang">Void
+cloneTableSchema(TableNametableName,
+TableNamenewTableName,
+booleanpreserveSplits)
+Create a new table by cloning the existent table
schema.
+
+
+
default https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
title="class or interface in java.lang">Void
compact(TableNametableName)
Compact a table.
-
+
default https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
title="class or interface in java.lang">Void
compact(TableNametableName,
byte[]columnFamily)
Compact a column family within a table.
-
+
https://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true;
title="class or interface in java.util.concurrent">CompletableFuturehttps://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true;
title="class or interface in java.lang">Void
compact(TableNametableName,
byte[]columnFamily,
@@ -247,40 +255,40 @@ public interface Compact a column family within a table.
-
+
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html
b/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html
index 57aa892..96d3860 100644
---
a/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html
+++
b/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.DeleteCommand.html
@@ -162,8 +162,8 @@ extends Constructor and Description
-DeleteCommand(org.apache.hadoop.conf.Configurationconf,
- org.apache.commons.cli.CommandLinecmdline)
+DeleteCommand(org.apache.hadoop.conf.Configurationconf,
+
org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLinecmdline)
@@ -230,14 +230,14 @@ extends
+
DeleteCommand
DeleteCommand(org.apache.hadoop.conf.Configurationconf,
- org.apache.commons.cli.CommandLinecmdline)
+
org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLinecmdline)
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.DescribeCommand.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.DescribeCommand.html
b/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.DescribeCommand.html
index 5de92de..dcfda4e 100644
---
a/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.DescribeCommand.html
+++
b/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.DescribeCommand.html
@@ -162,8 +162,8 @@ extends Constructor and Description
-DescribeCommand(org.apache.hadoop.conf.Configurationconf,
-
org.apache.commons.cli.CommandLinecmdline)
+DescribeCommand(org.apache.hadoop.conf.Configurationconf,
+
org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLinecmdline)
@@ -224,14 +224,14 @@ extends
+
DescribeCommand
DescribeCommand(org.apache.hadoop.conf.Configurationconf,
-org.apache.commons.cli.CommandLinecmdline)
+
org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLinecmdline)
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.HelpCommand.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.HelpCommand.html
b/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.HelpCommand.html
index b9a6076..769fc54 100644
---
a/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.HelpCommand.html
+++
b/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.HelpCommand.html
@@ -162,8 +162,8 @@ extends Constructor and Description
-HelpCommand(org.apache.hadoop.conf.Configurationconf,
- org.apache.commons.cli.CommandLinecmdline)
+HelpCommand(org.apache.hadoop.conf.Configurationconf,
+
org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLinecmdline)
@@ -224,14 +224,14 @@ extends
+
HelpCommand
HelpCommand(org.apache.hadoop.conf.Configurationconf,
-org.apache.commons.cli.CommandLinecmdline)
+
org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLinecmdline)
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.HistoryCommand.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.HistoryCommand.html
b/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.HistoryCommand.html
index d653bda..89762c6 100644
---
a/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.HistoryCommand.html
+++
b/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.HistoryCommand.html
@@ -173,8 +173,8 @@ extends Constructor and Description
-HistoryCommand(org.apache.hadoop.conf.Configurationconf,
-
org.apache.commons.cli.CommandLinecmdline)
+HistoryCommand(org.apache.hadoop.conf.Configurationconf,
+
org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLinecmdline)
@@ -272,14 +272,14 @@ extends
+
HistoryCommand
HistoryCommand(org.apache.hadoop.conf.Configurationconf,
- org.apache.commons.cli.CommandLinecmdline)
+
org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLinecmdline)
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupCommands.MergeCommand.html
--
diff
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/16541468/apidocs/org/apache/hadoop/hbase/client/Scan.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/client/Scan.html
b/apidocs/org/apache/hadoop/hbase/client/Scan.html
index 1ca681e..b93a975 100644
--- a/apidocs/org/apache/hadoop/hbase/client/Scan.html
+++ b/apidocs/org/apache/hadoop/hbase/client/Scan.html
@@ -18,7 +18,7 @@
catch(err) {
}
//-->
-var methods =
{"i0":10,"i1":10,"i2":9,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":42,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":42,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":42,"i55":42,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":42,"i67":42,"i68":42,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10};
+var methods =
{"i0":10,"i1":10,"i2":9,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":42,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":42,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":42,"i55":42,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":42,"i67":42,"i68":42,"i69":10,"i70":10,"i71":42,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10};
var tabs = {65535:["t0","All Methods"],1:["t1","Static
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete
Methods"],32:["t6","Deprecated Methods"]};
var altColor = "altColor";
var rowColor = "rowColor";
@@ -147,7 +147,7 @@ extends setTimeRange.
- To only retrieve columns with a specific timestamp, call setTimestamp
+ To only retrieve columns with a specific timestamp, call setTimestamp
.
To limit the number of versions of each column to be returned, call setMaxVersions.
@@ -723,37 +723,46 @@ extends
Scan
-setTimeStamp(longtimestamp)
+setTimestamp(longtimestamp)
Get versions of columns with the specified timestamp.
+Scan
+setTimeStamp(longtimestamp)
+Deprecated.
+As of release 2.0.0, this
will be removed in HBase 3.0.0.
+ Use setTimestamp(long)
instead
+
+
+
+
https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
toMap(intmaxCols)
Compile the details beyond the scope of getFingerprint
(row, columns,
timestamps, etc.) into a Map along with the fingerprinted information.
-
+
Scan
withStartRow(byte[]startRow)
Set the start row of the scan.
-
+
Scan
withStartRow(byte[]startRow,
booleaninclusive)
Set the start row of the scan.
-
+
Scan
withStopRow(byte[]stopRow)
Set the stop row of the scan.
-
+
Scan
withStopRow(byte[]stopRow,
booleaninclusive)
@@ -1064,8 +1073,11 @@ public
setTimeStamp
-publicScansetTimeStamp(longtimestamp)
- throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
+https://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true;
title="class or interface in java.lang">@Deprecated
+publicScansetTimeStamp(longtimestamp)
+ throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
+Deprecated.As of release 2.0.0, this will be removed in HBase
3.0.0.
+ Use setTimestamp(long)
instead
Get versions of columns with the specified timestamp. Note,
default maximum
versions to return is 1. If your time range spans more than one version
and you want all versions returned, up the number of versions beyond the
@@ -1083,13 +1095,35 @@ public
+
+
+
+
+setTimestamp
+publicScansetTimestamp(longtimestamp)
+Get versions of columns with the specified timestamp. Note,
default maximum
+ versions to return is 1. If your time range spans more than one version
+ and you want all versions returned, up the number of versions beyond the
+ defaut.
+
+Parameters:
+timestamp - version timestamp
+Returns:
+this
+See Also:
+setMaxVersions(),
+setMaxVersions(int)
+
+
+
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f1ebf5b6/apidocs/overview-tree.html
--
diff --git a/apidocs/overview-tree.html b/apidocs/overview-tree.html
index 72a723d..d331c6f 100644
--- a/apidocs/overview-tree.html
+++ b/apidocs/overview-tree.html
@@ -891,32 +891,32 @@
java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
title="class or interface in java.io">Serializable)
org.apache.hadoop.hbase.util.Order
-org.apache.hadoop.hbase.KeepDeletedCells
org.apache.hadoop.hbase.MemoryCompactionPolicy
+org.apache.hadoop.hbase.KeepDeletedCells
org.apache.hadoop.hbase.CompareOperator
org.apache.hadoop.hbase.CellBuilderType
-org.apache.hadoop.hbase.filter.BitComparator.BitwiseOp
org.apache.hadoop.hbase.filter.FilterList.Operator
org.apache.hadoop.hbase.filter.CompareFilter.CompareOp
-org.apache.hadoop.hbase.filter.Filter.ReturnCode
+org.apache.hadoop.hbase.filter.BitComparator.BitwiseOp
org.apache.hadoop.hbase.filter.RegexStringComparator.EngineType
+org.apache.hadoop.hbase.filter.Filter.ReturnCode
org.apache.hadoop.hbase.io.encoding.DataBlockEncoding
org.apache.hadoop.hbase.regionserver.BloomType
+org.apache.hadoop.hbase.quotas.SpaceViolationPolicy
org.apache.hadoop.hbase.quotas.ThrottlingException.Type
org.apache.hadoop.hbase.quotas.QuotaScope
-org.apache.hadoop.hbase.quotas.ThrottleType
org.apache.hadoop.hbase.quotas.QuotaType
-org.apache.hadoop.hbase.quotas.SpaceViolationPolicy
-org.apache.hadoop.hbase.client.Durability
+org.apache.hadoop.hbase.quotas.ThrottleType
org.apache.hadoop.hbase.client.SnapshotType
-org.apache.hadoop.hbase.client.MasterSwitchType
-org.apache.hadoop.hbase.client.CompactType
+org.apache.hadoop.hbase.client.Durability
org.apache.hadoop.hbase.client.MobCompactPartitionPolicy
-org.apache.hadoop.hbase.client.CompactionState
-org.apache.hadoop.hbase.client.Scan.ReadType
-org.apache.hadoop.hbase.client.RequestController.ReturnCode
org.apache.hadoop.hbase.client.IsolationLevel
+org.apache.hadoop.hbase.client.RequestController.ReturnCode
+org.apache.hadoop.hbase.client.Scan.ReadType
+org.apache.hadoop.hbase.client.CompactionState
+org.apache.hadoop.hbase.client.MasterSwitchType
org.apache.hadoop.hbase.client.Consistency
+org.apache.hadoop.hbase.client.CompactType
org.apache.hadoop.hbase.client.security.SecurityCapability
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f1ebf5b6/book.html
--
diff --git a/book.html b/book.html
index 79cba7d..80bdf26 100644
--- a/book.html
+++ b/book.html
@@ -6610,12 +6610,56 @@ Quitting...
13.
Upgrade Paths
-13.1. Upgrading
from 0.98.x to 1.x
+13.1. Upgrading
to 1.4+
+
+13.1.1. Replication
peers TableCFs config
+
+Before 1.4, the table name cant include namespace for replication
peers TableCFs config. It was fixed by add TableCFs to
ReplicationPeerConfig which was stored on Zookeeper. So when upgrade to 1.4,
you have to update the original ReplicationPeerConfig data on Zookeeper
firstly. There are four steps to upgrade when your cluster have a replication
peer with TableCFs config.
+
+
+
+
+Disable the replication peer.
+
+
+If master has permission to write replication peer znode, then rolling
update master directly. If not, use TableCFsUpdater tool to update the
replication peers config.
+
+
+
+
+
+$ bin/hbase org.apache.hadoop.hbase.replication.master.TableCFsUpdater
update
+
+
+
+
+
+Rolling update regionservers.
+
+
+Enable the replication peer.
+
+
+
+
+Notes:
+
+
+
+
+Cant use the old client(before 1.4) to change the replication
peers config. Because the client will write config to Zookeeper
directly, the old client will miss TableCFs config. And the old client write
TableCFs config to the old tablecfs znode, it will not work for new version
regionserver.
+
+
+
+
+
+
+13.2. Upgrading
from 0.98.x to 1.x
In this section we first note the significant changes that come in with
1.0.0+ HBase and then we go over the upgrade process. Be sure to read the
significant changes section with care so you avoid surprises.
-13.1.1. Changes of Note!
+13.2.1. Changes of Note!
In here we list important changes that are in 1.0.0+ since 0.98.x., changes
you should be aware that will go into effect once you upgrade.
@@ -6657,7 +6701,7 @@ using 0.98.11 servers with any other client version.
-13.1.2. Rolling upgrade from 0.98.x to
HBase 1.0.0
+13.2.2. Rolling upgrade from 0.98.x to
HBase 1.0.0
@@ -6676,7 +6720,7 @@ You cannot do a rolling
upgrade from 0.96.x
-13.1.3. Scanner Caching has Changed
+13.2.3. Scanner
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4dc2a2e8/apidocs/overview-tree.html
--
diff --git a/apidocs/overview-tree.html b/apidocs/overview-tree.html
index d331c6f..72a723d 100644
--- a/apidocs/overview-tree.html
+++ b/apidocs/overview-tree.html
@@ -891,32 +891,32 @@
java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
title="class or interface in java.io">Serializable)
org.apache.hadoop.hbase.util.Order
-org.apache.hadoop.hbase.MemoryCompactionPolicy
org.apache.hadoop.hbase.KeepDeletedCells
+org.apache.hadoop.hbase.MemoryCompactionPolicy
org.apache.hadoop.hbase.CompareOperator
org.apache.hadoop.hbase.CellBuilderType
+org.apache.hadoop.hbase.filter.BitComparator.BitwiseOp
org.apache.hadoop.hbase.filter.FilterList.Operator
org.apache.hadoop.hbase.filter.CompareFilter.CompareOp
-org.apache.hadoop.hbase.filter.BitComparator.BitwiseOp
-org.apache.hadoop.hbase.filter.RegexStringComparator.EngineType
org.apache.hadoop.hbase.filter.Filter.ReturnCode
+org.apache.hadoop.hbase.filter.RegexStringComparator.EngineType
org.apache.hadoop.hbase.io.encoding.DataBlockEncoding
org.apache.hadoop.hbase.regionserver.BloomType
-org.apache.hadoop.hbase.quotas.SpaceViolationPolicy
org.apache.hadoop.hbase.quotas.ThrottlingException.Type
org.apache.hadoop.hbase.quotas.QuotaScope
-org.apache.hadoop.hbase.quotas.QuotaType
org.apache.hadoop.hbase.quotas.ThrottleType
-org.apache.hadoop.hbase.client.SnapshotType
+org.apache.hadoop.hbase.quotas.QuotaType
+org.apache.hadoop.hbase.quotas.SpaceViolationPolicy
org.apache.hadoop.hbase.client.Durability
+org.apache.hadoop.hbase.client.SnapshotType
+org.apache.hadoop.hbase.client.MasterSwitchType
+org.apache.hadoop.hbase.client.CompactType
org.apache.hadoop.hbase.client.MobCompactPartitionPolicy
-org.apache.hadoop.hbase.client.IsolationLevel
-org.apache.hadoop.hbase.client.RequestController.ReturnCode
-org.apache.hadoop.hbase.client.Scan.ReadType
org.apache.hadoop.hbase.client.CompactionState
-org.apache.hadoop.hbase.client.MasterSwitchType
+org.apache.hadoop.hbase.client.Scan.ReadType
+org.apache.hadoop.hbase.client.RequestController.ReturnCode
+org.apache.hadoop.hbase.client.IsolationLevel
org.apache.hadoop.hbase.client.Consistency
-org.apache.hadoop.hbase.client.CompactType
org.apache.hadoop.hbase.client.security.SecurityCapability
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4dc2a2e8/apidocs/src-html/org/apache/hadoop/hbase/LocalHBaseCluster.html
--
diff --git a/apidocs/src-html/org/apache/hadoop/hbase/LocalHBaseCluster.html
b/apidocs/src-html/org/apache/hadoop/hbase/LocalHBaseCluster.html
index a0ae102..1d2d704 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/LocalHBaseCluster.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/LocalHBaseCluster.html
@@ -149,7 +149,7 @@
141
142// Always have masters and
regionservers come up on port '0' so we don't
143// clash over default ports.
-144if
(conf.getBoolean(ASSIGN_RANDOM_PORTS, true)) {
+144if
(conf.getBoolean(ASSIGN_RANDOM_PORTS, false)) {
145 conf.set(HConstants.MASTER_PORT,
"0");
146
conf.set(HConstants.REGIONSERVER_PORT, "0");
147 if
(conf.getInt(HConstants.REGIONSERVER_INFO_PORT, 0) != -1) {
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4dc2a2e8/book.html
--
diff --git a/book.html b/book.html
index 27131f0..79cba7d 100644
--- a/book.html
+++ b/book.html
@@ -5203,6 +5203,18 @@ config.set(
+
+7.6.
Timeout settings
+
+HBase provides many timeout settings to limit the execution time of
different remote operations.
+
+
+The hbase.rpc.timeout property limits how long an RPC call can
run before it times out.
+You can also specify a timeout for read and write operations using
hbase.rpc.read.timeout and hbase.rpc.write.timeout
configuration properties. In the absence of these properties
hbase.rpc.timeout will be used.
+A higher-level timeout is hbase.client.operation.timeout which is
valid for each client call.
+Timeout for scan operations is controlled differently. To set it you can use
hbase.client.scanner.timeout.period property.
+
+
@@ -37288,7 +37300,7 @@ The server will return cellblocks compressed using this
same compressor as long
Version 3.0.0-SNAPSHOT
-Last updated 2018-03-21 14:29:53 UTC
+Last updated 2018-03-22 14:29:40 UTC
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4dc2a2e8/bulk-loads.html
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6eb695c8/devapidocs/org/apache/hadoop/hbase/regionserver/wal/FSHLog.RingBufferEventHandler.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/regionserver/wal/FSHLog.RingBufferEventHandler.html
b/devapidocs/org/apache/hadoop/hbase/regionserver/wal/FSHLog.RingBufferEventHandler.html
index 92de9a1..a544da6 100644
---
a/devapidocs/org/apache/hadoop/hbase/regionserver/wal/FSHLog.RingBufferEventHandler.html
+++
b/devapidocs/org/apache/hadoop/hbase/regionserver/wal/FSHLog.RingBufferEventHandler.html
@@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
-class FSHLog.RingBufferEventHandler
+class FSHLog.RingBufferEventHandler
extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
implements com.lmax.disruptor.EventHandlerRingBufferTruck,
com.lmax.disruptor.LifecycleAware
Handler that is run by the disruptor ringbuffer consumer.
Consumer is a SINGLE
@@ -302,7 +302,7 @@ implements com.lmax.disruptor.EventHandler
syncRunners
-private finalFSHLog.SyncRunner[] syncRunners
+private finalFSHLog.SyncRunner[] syncRunners
@@ -311,7 +311,7 @@ implements com.lmax.disruptor.EventHandler
syncFutures
-private finalSyncFuture[] syncFutures
+private finalSyncFuture[] syncFutures
@@ -320,7 +320,7 @@ implements com.lmax.disruptor.EventHandler
syncFuturesCount
-privatehttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicInteger.html?is-external=true;
title="class or interface in java.util.concurrent.atomic">AtomicInteger syncFuturesCount
+privatehttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicInteger.html?is-external=true;
title="class or interface in java.util.concurrent.atomic">AtomicInteger syncFuturesCount
@@ -329,7 +329,7 @@ implements com.lmax.disruptor.EventHandler
zigzagLatch
-private volatileFSHLog.SafePointZigZagLatch zigzagLatch
+private volatileFSHLog.SafePointZigZagLatch zigzagLatch
@@ -338,7 +338,7 @@ implements com.lmax.disruptor.EventHandler
exception
-privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
title="class or interface in java.lang">Exception exception
+privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
title="class or interface in java.lang">Exception exception
Set if we get an exception appending or syncing so that all
subsequence appends and syncs on
this WAL fail until WAL is replaced.
@@ -349,7 +349,7 @@ implements com.lmax.disruptor.EventHandler
safePointWaiter
-private finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object safePointWaiter
+private finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object safePointWaiter
Object to block on while waiting on safe point.
@@ -359,7 +359,7 @@ implements com.lmax.disruptor.EventHandler
shutdown
-private volatileboolean shutdown
+private volatileboolean shutdown
@@ -368,7 +368,7 @@ implements com.lmax.disruptor.EventHandler
syncRunnerIndex
-privateint syncRunnerIndex
+privateint syncRunnerIndex
Which syncrunner to use next.
@@ -386,7 +386,7 @@ implements com.lmax.disruptor.EventHandler
RingBufferEventHandler
-RingBufferEventHandler(intsyncRunnerCount,
+RingBufferEventHandler(intsyncRunnerCount,
intmaxHandlersCount)
@@ -404,7 +404,7 @@ implements com.lmax.disruptor.EventHandler
cleanupOutstandingSyncsOnException
-privatevoidcleanupOutstandingSyncsOnException(longsequence,
+privatevoidcleanupOutstandingSyncsOnException(longsequence,
https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
title="class or interface in java.lang">Exceptione)
@@ -414,7 +414,7 @@ implements com.lmax.disruptor.EventHandler
isOutstandingSyncs
-privatebooleanisOutstandingSyncs()
+privatebooleanisOutstandingSyncs()
Returns:
True if outstanding sync futures still
@@ -427,7 +427,7 @@ implements com.lmax.disruptor.EventHandler
isOutstandingSyncsFromRunners
-privatebooleanisOutstandingSyncsFromRunners()
+privatebooleanisOutstandingSyncsFromRunners()
@@ -436,7 +436,7 @@ implements com.lmax.disruptor.EventHandler
onEvent
-publicvoidonEvent(RingBufferTrucktruck,
+publicvoidonEvent(RingBufferTrucktruck,
longsequence,
booleanendOfBatch)
throws https://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
title="class or interface in java.lang">Exception
@@ -454,7 +454,7 @@ implements com.lmax.disruptor.EventHandler
attainSafePoint
-FSHLog.SafePointZigZagLatchattainSafePoint()
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bd675fa3/devapidocs/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineTableProcedure.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineTableProcedure.html
b/devapidocs/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineTableProcedure.html
index bd28f25..1e2eb5b 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineTableProcedure.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/procedure/AbstractStateMachineTableProcedure.html
@@ -18,8 +18,8 @@
catch(err) {
}
//-->
-var methods =
{"i0":10,"i1":10,"i2":10,"i3":6,"i4":6,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10};
-var tabs = {65535:["t0","All Methods"],2:["t2","Instance
Methods"],4:["t3","Abstract Methods"],8:["t4","Concrete Methods"]};
+var methods =
{"i0":10,"i1":9,"i2":10,"i3":10,"i4":6,"i5":6,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10};
+var tabs = {65535:["t0","All Methods"],1:["t1","Static
Methods"],2:["t2","Instance Methods"],4:["t3","Abstract
Methods"],8:["t4","Concrete Methods"]};
var altColor = "altColor";
var rowColor = "rowColor";
var tableTab = "tableTab";
@@ -128,7 +128,7 @@ var activeTableTab = "activeTableTab";
@InterfaceAudience.Private
-public abstract class AbstractStateMachineTableProcedureTState
+public abstract class AbstractStateMachineTableProcedureTState
extends StateMachineProcedureMasterProcedureEnv,TState
implements TableProcedureInterface
Base class for all the Table procedures that want to use a
StateMachineProcedure.
@@ -233,7 +233,7 @@ implements
-All MethodsInstance MethodsAbstract MethodsConcrete Methods
+All MethodsStatic MethodsInstance MethodsAbstract MethodsConcrete Methods
Modifier and Type
Method and Description
@@ -245,52 +245,59 @@ implements
+protected static void
+checkOnline(MasterProcedureEnvenv,
+ RegionInfori)
+Check region is online.
+
+
+
protected void
checkTableModifiable(MasterProcedureEnvenv)
Check whether a table is modifiable - exists and either
offline or online with config set
-
+
protected org.apache.hadoop.fs.Path
getRegionDir(MasterProcedureEnvenv,
RegionInforegion)
-
+
abstract TableName
getTableName()
-
+
abstract TableProcedureInterface.TableOperationType
getTableOperationType()
Given an operation type we can take decisions about what to
do with pending operations.
-
+
protected User
getUser()
-
+
protected void
preflightChecks(MasterProcedureEnvenv,
https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
title="class or interface in java.lang">Booleanenabled)
Check that cluster is up and master is running.
-
+
protected void
releaseLock(MasterProcedureEnvenv)
The user should override this method, and release lock if
necessary.
-
+
protected void
releaseSyncLatch()
-
+
protected void
setUser(Useruser)
-
+
void
toStringClassDetails(https://docs.oracle.com/javase/8/docs/api/java/lang/StringBuilder.html?is-external=true;
title="class or interface in java.lang">StringBuildersb)
Extend the toString() information with the procedure details
@@ -339,7 +346,7 @@ implements
syncLatch
-private finalProcedurePrepareLatch syncLatch
+private finalProcedurePrepareLatch syncLatch
@@ -348,7 +355,7 @@ implements
user
-privateUser user
+privateUser user
@@ -365,7 +372,7 @@ implements
AbstractStateMachineTableProcedure
-protectedAbstractStateMachineTableProcedure()
+protectedAbstractStateMachineTableProcedure()
@@ -374,7 +381,7 @@ implements
AbstractStateMachineTableProcedure
-protectedAbstractStateMachineTableProcedure(MasterProcedureEnvenv)
+protectedAbstractStateMachineTableProcedure(MasterProcedureEnvenv)
@@ -383,7 +390,7 @@ implements
AbstractStateMachineTableProcedure
-protectedAbstractStateMachineTableProcedure(MasterProcedureEnvenv,
+protectedAbstractStateMachineTableProcedure(MasterProcedureEnvenv,
ProcedurePrepareLatchlatch)
Parameters:
@@ -405,7 +412,7 @@ implements
getTableName
-public abstractTableNamegetTableName()
+public abstractTableNamegetTableName()
Specified by:
getTableNamein
interfaceTableProcedureInterface
@@ -420,7 +427,7 @@ implements
getTableOperationType
-public abstractTableProcedureInterface.TableOperationTypegetTableOperationType()
+public abstractTableProcedureInterface.TableOperationTypegetTableOperationType()
Description copied from
interface:TableProcedureInterface
Given an operation type we can take decisions about what to
do with pending operations.
e.g. if we get a delete and we have some table operation pending (e.g. add
column)
@@ -439,7 +446,7 @@ implements
toStringClassDetails
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8b1eaec1/apidocs/src-html/org/apache/hadoop/hbase/client/CoprocessorDescriptor.html
--
diff --git
a/apidocs/src-html/org/apache/hadoop/hbase/client/CoprocessorDescriptor.html
b/apidocs/src-html/org/apache/hadoop/hbase/client/CoprocessorDescriptor.html
new file mode 100644
index 000..1064470
--- /dev/null
+++ b/apidocs/src-html/org/apache/hadoop/hbase/client/CoprocessorDescriptor.html
@@ -0,0 +1,123 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+Source code
+
+
+
+
+001/**
+002 *
+003 * Licensed to the Apache Software
Foundation (ASF) under one
+004 * or more contributor license
agreements. See the NOTICE file
+005 * distributed with this work for
additional information
+006 * regarding copyright ownership. The
ASF licenses this file
+007 * to you under the Apache License,
Version 2.0 (the
+008 * "License"); you may not use this file
except in compliance
+009 * with the License. You may obtain a
copy of the License at
+010 *
+011 *
http://www.apache.org/licenses/LICENSE-2.0
+012 *
+013 * Unless required by applicable law or
agreed to in writing, software
+014 * distributed under the License is
distributed on an "AS IS" BASIS,
+015 * WITHOUT WARRANTIES OR CONDITIONS OF
ANY KIND, either express or implied.
+016 * See the License for the specific
language governing permissions and
+017 * limitations under the License.
+018 */
+019package org.apache.hadoop.hbase.client;
+020
+021import java.util.Map;
+022import java.util.Optional;
+023import
org.apache.yetus.audience.InterfaceAudience;
+024
+025/**
+026 * CoprocessorDescriptor contains the
details about how to build a coprocessor.
+027 * This class is a pojo so there are no
checks for the details carried by this class.
+028 * Use {@link
CoprocessorDescriptorBuilder} to instantiate a CoprocessorDescriptor
+029 */
+030@InterfaceAudience.Public
+031public interface CoprocessorDescriptor
{
+032 /**
+033 * @return the name of the class or
interface represented by this object.
+034 */
+035 String getClassName();
+036
+037 /**
+038 * @return Path of the jar file. If
it's null, the class will be loaded from default classloader.
+039 */
+040 OptionalString getJarPath();
+041
+042 /**
+043 * @return The order to execute this
coprocessor
+044 */
+045 int getPriority();
+046
+047 /**
+048 * @return Arbitrary key-value
parameter pairs passed into the coprocessor.
+049 */
+050 MapString, String
getProperties();
+051}
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8b1eaec1/apidocs/src-html/org/apache/hadoop/hbase/client/CoprocessorDescriptorBuilder.html
--
diff --git
a/apidocs/src-html/org/apache/hadoop/hbase/client/CoprocessorDescriptorBuilder.html
b/apidocs/src-html/org/apache/hadoop/hbase/client/CoprocessorDescriptorBuilder.html
new file mode 100644
index 000..ef0d80f
--- /dev/null
+++
b/apidocs/src-html/org/apache/hadoop/hbase/client/CoprocessorDescriptorBuilder.html
@@ -0,0 +1,190 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+Source code
+
+
+
+
+001/**
+002 *
+003 * Licensed to the Apache Software
Foundation (ASF) under one
+004 * or more contributor license
agreements. See the NOTICE file
+005 * distributed with this work for
additional information
+006 * regarding copyright ownership. The
ASF licenses this file
+007 * to you under the Apache License,
Version 2.0 (the
+008 * "License"); you may not use this file
except in compliance
+009 * with the License. You may obtain a
copy of the License at
+010 *
+011 *
http://www.apache.org/licenses/LICENSE-2.0
+012 *
+013 * Unless required by applicable law or
agreed to in writing, software
+014 * distributed under the License is
distributed on an "AS IS" BASIS,
+015 * WITHOUT WARRANTIES OR CONDITIONS OF
ANY KIND, either express or implied.
+016 * See the License for the specific
language governing permissions and
+017 * limitations under the License.
+018 */
+019package org.apache.hadoop.hbase.client;
+020
+021import java.util.Collections;
+022import java.util.Map;
+023import java.util.Objects;
+024import java.util.Optional;
+025import java.util.TreeMap;
+026import
org.apache.hadoop.hbase.Coprocessor;
+027import
org.apache.yetus.audience.InterfaceAudience;
+028
+029/**
+030 * Used to build the {@link
CoprocessorDescriptor}
+031 */
+032@InterfaceAudience.Public
+033public final class
CoprocessorDescriptorBuilder {
+034
+035 public static CoprocessorDescriptor
of(String className) {
+036return new
CoprocessorDescriptorBuilder(className).build();
+037 }
+038
+039 public static
CoprocessorDescriptorBuilder newBuilder(String className) {
+040return new
CoprocessorDescriptorBuilder(className);
+041 }
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a754d895/devapidocs/org/apache/hadoop/hbase/MetaTableAccessor.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/MetaTableAccessor.html
b/devapidocs/org/apache/hadoop/hbase/MetaTableAccessor.html
index 998311f..4c72218 100644
--- a/devapidocs/org/apache/hadoop/hbase/MetaTableAccessor.html
+++ b/devapidocs/org/apache/hadoop/hbase/MetaTableAccessor.html
@@ -667,7 +667,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
getReplicationBarrierResult(Resultresult)
-private static long[]
+static long[]
getReplicationBarriers(Resultresult)
@@ -2096,7 +2096,7 @@ public static
getTableStates
-public statichttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
title="class or interface in java.util">MapTableName,TableStategetTableStates(Connectionconn)
+public statichttps://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
title="class or interface in java.util">MapTableName,TableStategetTableStates(Connectionconn)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
Fetch table states from META table
@@ -2115,7 +2115,7 @@ public static
updateTableState
-public staticvoidupdateTableState(Connectionconn,
+public staticvoidupdateTableState(Connectionconn,
TableNametableName,
TableState.Stateactual)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
@@ -2136,7 +2136,7 @@ public static
getTableState
@Nullable
-public staticTableStategetTableState(Resultr)
+public staticTableStategetTableState(Resultr)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
Decode table state from META Result.
Should contain cell from HConstants.TABLE_FAMILY
@@ -2156,7 +2156,7 @@ public static
getRegionCount
-public staticintgetRegionCount(org.apache.hadoop.conf.Configurationc,
+public staticintgetRegionCount(org.apache.hadoop.conf.Configurationc,
TableNametableName)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
Count regions in hbase:meta for passed
table.
@@ -2177,7 +2177,7 @@ public static
getRegionCount
-public staticintgetRegionCount(Connectionconnection,
+public staticintgetRegionCount(Connectionconnection,
TableNametableName)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
Count regions in hbase:meta for passed
table.
@@ -2198,7 +2198,7 @@ public static
makePutFromRegionInfo
-public staticPutmakePutFromRegionInfo(RegionInforegionInfo,
+public staticPutmakePutFromRegionInfo(RegionInforegionInfo,
longts)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
Generates and returns a Put containing the region into for
the catalog table
@@ -2214,7 +2214,7 @@ public static
makeDeleteFromRegionInfo
-private staticDeletemakeDeleteFromRegionInfo(RegionInforegionInfo,
+private staticDeletemakeDeleteFromRegionInfo(RegionInforegionInfo,
longts)
Generates and returns a Delete containing the region info
for the catalog
table
@@ -2226,7 +2226,7 @@ public static
addDaughtersToPut
-public staticPutaddDaughtersToPut(Putput,
+public staticPutaddDaughtersToPut(Putput,
RegionInfosplitA,
RegionInfosplitB)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
@@ -2243,7 +2243,7 @@ public static
putToMetaTable
-private staticvoidputToMetaTable(Connectionconnection,
+private staticvoidputToMetaTable(Connectionconnection,
Putp)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
Put the passed p to the
hbase:meta table.
@@ -2262,7 +2262,7 @@ public static
put
-private staticvoidput(Tablet,
+private staticvoidput(Tablet,
Putp)
throws
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/apidocs/overview-tree.html
--
diff --git a/apidocs/overview-tree.html b/apidocs/overview-tree.html
index 030c910..0e9aafb 100644
--- a/apidocs/overview-tree.html
+++ b/apidocs/overview-tree.html
@@ -889,33 +889,33 @@
java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
title="class or interface in java.io">Serializable)
org.apache.hadoop.hbase.util.Order
-org.apache.hadoop.hbase.KeepDeletedCells
org.apache.hadoop.hbase.MemoryCompactionPolicy
+org.apache.hadoop.hbase.KeepDeletedCells
org.apache.hadoop.hbase.CompareOperator
org.apache.hadoop.hbase.ProcedureState
org.apache.hadoop.hbase.CellBuilderType
-org.apache.hadoop.hbase.filter.BitComparator.BitwiseOp
org.apache.hadoop.hbase.filter.FilterList.Operator
org.apache.hadoop.hbase.filter.CompareFilter.CompareOp
-org.apache.hadoop.hbase.filter.Filter.ReturnCode
+org.apache.hadoop.hbase.filter.BitComparator.BitwiseOp
org.apache.hadoop.hbase.filter.RegexStringComparator.EngineType
+org.apache.hadoop.hbase.filter.Filter.ReturnCode
org.apache.hadoop.hbase.io.encoding.DataBlockEncoding
org.apache.hadoop.hbase.regionserver.BloomType
+org.apache.hadoop.hbase.quotas.SpaceViolationPolicy
org.apache.hadoop.hbase.quotas.ThrottlingException.Type
org.apache.hadoop.hbase.quotas.QuotaScope
-org.apache.hadoop.hbase.quotas.ThrottleType
org.apache.hadoop.hbase.quotas.QuotaType
-org.apache.hadoop.hbase.quotas.SpaceViolationPolicy
-org.apache.hadoop.hbase.client.Durability
+org.apache.hadoop.hbase.quotas.ThrottleType
org.apache.hadoop.hbase.client.SnapshotType
-org.apache.hadoop.hbase.client.MasterSwitchType
-org.apache.hadoop.hbase.client.CompactType
+org.apache.hadoop.hbase.client.Durability
org.apache.hadoop.hbase.client.MobCompactPartitionPolicy
-org.apache.hadoop.hbase.client.CompactionState
-org.apache.hadoop.hbase.client.Scan.ReadType
-org.apache.hadoop.hbase.client.RequestController.ReturnCode
org.apache.hadoop.hbase.client.IsolationLevel
+org.apache.hadoop.hbase.client.RequestController.ReturnCode
+org.apache.hadoop.hbase.client.Scan.ReadType
+org.apache.hadoop.hbase.client.CompactionState
+org.apache.hadoop.hbase.client.MasterSwitchType
org.apache.hadoop.hbase.client.Consistency
+org.apache.hadoop.hbase.client.CompactType
org.apache.hadoop.hbase.client.security.SecurityCapability
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/book.html
--
diff --git a/book.html b/book.html
index 4ee0401..7373777 100644
--- a/book.html
+++ b/book.html
@@ -1448,11 +1448,6 @@ For most configurations, a restart is needed for servers
to pick up changes. Dyn
yes
yes
-
-1.1
-yes
-Running
with JDK 8 will work but is not well tested.
-
@@ -1612,16 +1607,14 @@ earlier versions of Hadoop. See the table below for
requirements specific to dif
-
-
-
-
-
+
+
+
+
-HBase-1.1.x
HBase-1.2.x
HBase-1.3.x
HBase-2.0.x
@@ -1633,25 +1626,21 @@ earlier versions of Hadoop. See the table below for
requirements specific to dif
X
X
X
-X
Hadoop-2.1.0-beta
X
X
X
-X
Hadoop-2.2.0
-NT
X
X
X
Hadoop-2.3.x
-NT
X
X
X
@@ -1660,14 +1649,12 @@ earlier versions of Hadoop. See the table below for
requirements specific to dif
Hadoop-2.4.x
S
S
-S
X
Hadoop-2.5.x
S
S
-S
X
@@ -1675,11 +1662,9 @@ earlier versions of Hadoop. See the table below for
requirements specific to dif
X
X
X
-X
Hadoop-2.6.1+
-NT
S
S
S
@@ -1689,11 +1674,9 @@ earlier versions of Hadoop. See the table below for
requirements specific to dif
X
X
X
-X
Hadoop-2.7.1+
-NT
S
S
S
@@ -1703,21 +1686,18 @@ earlier versions of Hadoop. See the table below for
requirements specific to dif
X
X
X
-X
Hadoop-2.8.1
X
X
X
-X
Hadoop-3.0.0
NT
NT
NT
-NT
@@ -8092,8 +8072,10 @@ In particular:
Below we describe how the version dimension in HBase currently works.
See https://issues.apache.org/jira/browse/HBASE-2406;>HBASE-2406
for discussion of HBase versions. https://www.ngdata.com/bending-time-in-hbase/;>Bending time in HBase
makes for a good read on the version, or time, dimension in HBase.
-It has more detail on versioning than is provided here.
-As of this writing, the limitation Overwriting values at existing
timestamps mentioned in the article no longer holds in HBase.
+It has more detail on versioning than is provided here.
+
+
+As of this writing, the limitation Overwriting values at existing
timestamps mentioned in the article no longer holds in HBase.
This section is
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d347bde8/checkstyle.rss
--
diff --git a/checkstyle.rss b/checkstyle.rss
index ea6ab21..eab2875 100644
--- a/checkstyle.rss
+++ b/checkstyle.rss
@@ -25,8 +25,8 @@ under the License.
en-us
2007 - 2018 The Apache Software Foundation
- File: 3574,
- Errors: 16184,
+ File: 3579,
+ Errors: 16180,
Warnings: 0,
Infos: 0
@@ -1908,7 +1908,7 @@ under the License.
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.mob.MobUtils.java;>org/apache/hadoop/hbase/mob/MobUtils.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.DateTieredStoreEngine.java;>org/apache/hadoop/hbase/regionserver/DateTieredStoreEngine.java
0
@@ -1917,12 +1917,12 @@ under the License.
0
- 15
+ 1
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.DateTieredStoreEngine.java;>org/apache/hadoop/hbase/regionserver/DateTieredStoreEngine.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.mob.MobUtils.java;>org/apache/hadoop/hbase/mob/MobUtils.java
0
@@ -1931,7 +1931,7 @@ under the License.
0
- 1
+ 15
@@ -3784,7 +3784,7 @@ under the License.
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.quotas.TestQuotaObserverChore.java;>org/apache/hadoop/hbase/quotas/TestQuotaObserverChore.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.util.HFileArchiveTestingUtil.java;>org/apache/hadoop/hbase/util/HFileArchiveTestingUtil.java
0
@@ -3793,12 +3793,12 @@ under the License.
0
- 0
+ 9
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.util.HFileArchiveTestingUtil.java;>org/apache/hadoop/hbase/util/HFileArchiveTestingUtil.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.quotas.TestQuotaObserverChore.java;>org/apache/hadoop/hbase/quotas/TestQuotaObserverChore.java
0
@@ -3807,7 +3807,7 @@ under the License.
0
- 9
+ 0
@@ -5380,7 +5380,7 @@ under the License.
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.TestRegionSnapshotTask.java;>org/apache/hadoop/hbase/snapshot/TestRegionSnapshotTask.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.TestSwitchToStreamRead.java;>org/apache/hadoop/hbase/regionserver/TestSwitchToStreamRead.java
0
@@ -5394,7 +5394,7 @@ under the License.
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.TestRestoreSnapshotHelper.java;>org/apache/hadoop/hbase/snapshot/TestRestoreSnapshotHelper.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.TestRegionSnapshotTask.java;>org/apache/hadoop/hbase/snapshot/TestRegionSnapshotTask.java
0
@@ -5403,12 +5403,12 @@ under the License.
0
- 2
+ 0
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.TestSwitchToStreamRead.java;>org/apache/hadoop/hbase/regionserver/TestSwitchToStreamRead.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.TestRestoreSnapshotHelper.java;>org/apache/hadoop/hbase/snapshot/TestRestoreSnapshotHelper.java
0
@@ -5417,7 +5417,7 @@ under the License.
0
-
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6b94a2f2/devapidocs/org/apache/hadoop/hbase/HConstants.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/HConstants.html
b/devapidocs/org/apache/hadoop/hbase/HConstants.html
index 3949f2a..5d41c23 100644
--- a/devapidocs/org/apache/hadoop/hbase/HConstants.html
+++ b/devapidocs/org/apache/hadoop/hbase/HConstants.html
@@ -1486,6 +1486,18 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
REPLAY_QOS
+static byte[]
+REPLICATION_BARRIER_FAMILY
+The replication barrier family
+
+
+
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
+REPLICATION_BARRIER_FAMILY_STR
+The replication barrier family as a string
+
+
+
static boolean
REPLICATION_BULKLOAD_ENABLE_DEFAULT
@@ -1535,277 +1547,284 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
+static int
+REPLICATION_SCOPE_SERIAL
+Scope tag for serially scoped data
+ This data will be replicated to all peers by the order of sequence id.
+
+
+
static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
REPLICATION_SERVICE_CLASSNAME_DEFAULT
-
+
static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
REPLICATION_SINK_SERVICE_CLASSNAME
-
+
static int
REPLICATION_SOURCE_MAXTHREADS_DEFAULT
Maximum number of threads used by the replication source
for shipping edits to the sinks
-
+
static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
REPLICATION_SOURCE_MAXTHREADS_KEY
Maximum number of threads used by the replication source
for shipping edits to the sinks
-
+
static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
REPLICATION_SOURCE_SERVICE_CLASSNAME
-
+
static int
REPLICATION_SOURCE_TOTAL_BUFFER_DFAULT
-
+
static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
REPLICATION_SOURCE_TOTAL_BUFFER_KEY
Max total size of buffered entries in all replication
peers.
-
+
static int[]
RETRY_BACKOFF
Retrying we multiply hbase.client.pause setting by what we
have in this array until we
run out of array items.
-
+
static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
RPC_CODEC_CONF_KEY
Configuration key for setting RPC codec class name
-
+
static byte
RPC_CURRENT_VERSION
-
+
static byte[]
RPC_HEADER
The first four bytes of Hadoop RPC connections
-
+
static byte[]
SEQNUM_QUALIFIER
The open seqnum column qualifier
-
+
static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
SEQNUM_QUALIFIER_STR
The open seqnum column qualifier
-
+
static byte[]
SERVER_QUALIFIER
The server column qualifier
-
+
static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
SERVER_QUALIFIER_STR
The server column qualifier
-
+
static byte[]
SERVERNAME_QUALIFIER
-
+
static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
SERVERNAME_QUALIFIER_STR
The serverName column qualifier.
-
+
static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
SNAPSHOT_DIR_NAME
Name of the directory to store all snapshots.
-
+
static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
SNAPSHOT_RESTORE_FAILSAFE_NAME
-
+
static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
SNAPSHOT_RESTORE_TAKE_FAILSAFE_SNAPSHOT
-
+
static int
SOCKET_RETRY_WAIT_MS
The delay when re-trying a socket operation in a loop
(HBASE-4712)
-
+
static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
SPLIT_LOGDIR_NAME
Used to construct the name of the splitlog directory for a
region server
-
+
static byte[]
SPLITA_QUALIFIER
The lower-half split region column qualifier
-
+
static byte[]
SPLITB_QUALIFIER
The upper-half split region column qualifier
-
+
static byte[]
STARTCODE_QUALIFIER
The startcode column qualifier
-
+
static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb05e3e3/apidocs/org/apache/hadoop/hbase/HColumnDescriptor.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/HColumnDescriptor.html
b/apidocs/org/apache/hadoop/hbase/HColumnDescriptor.html
index 2a7df6f..b7f6e3e 100644
--- a/apidocs/org/apache/hadoop/hbase/HColumnDescriptor.html
+++ b/apidocs/org/apache/hadoop/hbase/HColumnDescriptor.html
@@ -97,7 +97,7 @@ var activeTableTab = "activeTableTab";
-http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">java.lang.Object
+https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">java.lang.Object
org.apache.hadoop.hbase.HColumnDescriptor
@@ -109,16 +109,16 @@ var activeTableTab = "activeTableTab";
All Implemented Interfaces:
-http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableHColumnDescriptor, ColumnFamilyDescriptor
+https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableHColumnDescriptor, ColumnFamilyDescriptor
Deprecated.
@InterfaceAudience.Public
- http://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true;
title="class or interface in java.lang">@Deprecated
+ https://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true;
title="class or interface in java.lang">@Deprecated
public class HColumnDescriptor
-extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
-implements ColumnFamilyDescriptor,
http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableHColumnDescriptor
+extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
+implements ColumnFamilyDescriptor,
https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableHColumnDescriptor
An HColumnDescriptor contains information about a column
family such as the
number of versions, compression settings, etc.
@@ -142,67 +142,67 @@ implements Field and Description
-static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
BLOCKCACHE
Deprecated.
-static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
BLOCKSIZE
Deprecated.
-static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
BLOOMFILTER
Deprecated.
-static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
CACHE_BLOOMS_ON_WRITE
Deprecated.
-static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
CACHE_DATA_IN_L1
Deprecated.
-static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
CACHE_DATA_ON_WRITE
Deprecated.
-static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
CACHE_INDEX_ON_WRITE
Deprecated.
-static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
COMPRESS_TAGS
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ec8bf761/checkstyle-aggregate.html
--
diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html
index 0328956..b705f85 100644
--- a/checkstyle-aggregate.html
+++ b/checkstyle-aggregate.html
@@ -7,7 +7,7 @@
-
+
Apache HBase Checkstyle Results
@@ -286,10 +286,10 @@
Warnings
Errors
-3544
+3550
0
0
-16537
+16524
Files
@@ -1259,9056 +1259,9051 @@
0
2
-org/apache/hadoop/hbase/client/ClusterConnection.java
-0
-0
-1
-
org/apache/hadoop/hbase/client/ColumnCountOnRowFilter.java
0
0
1
-
+
org/apache/hadoop/hbase/client/ColumnFamilyDescriptor.java
0
0
13
-
+
org/apache/hadoop/hbase/client/ColumnFamilyDescriptorBuilder.java
0
0
54
-
+
org/apache/hadoop/hbase/client/CompactType.java
0
0
2
-
+
org/apache/hadoop/hbase/client/CompleteScanResultCache.java
0
0
1
-
+
org/apache/hadoop/hbase/client/ConnectionFactory.java
0
0
3
-
+
org/apache/hadoop/hbase/client/ConnectionImplementation.java
0
0
9
-
+
org/apache/hadoop/hbase/client/ConnectionUtils.java
0
0
3
-
+
org/apache/hadoop/hbase/client/DelayingRunner.java
0
0
4
-
+
org/apache/hadoop/hbase/client/Delete.java
0
0
6
-
+
org/apache/hadoop/hbase/client/Get.java
0
0
10
-
+
org/apache/hadoop/hbase/client/HBaseAdmin.java
0
0
72
-
+
org/apache/hadoop/hbase/client/HConnectionTestingUtility.java
0
0
14
-
+
org/apache/hadoop/hbase/client/HRegionLocator.java
0
0
1
-
+
org/apache/hadoop/hbase/client/HTable.java
0
0
46
-
+
org/apache/hadoop/hbase/client/HTableMultiplexer.java
0
0
5
-
+
org/apache/hadoop/hbase/client/ImmutableHColumnDescriptor.java
0
0
2
-
+
org/apache/hadoop/hbase/client/ImmutableHRegionInfo.java
0
0
1
-
+
org/apache/hadoop/hbase/client/ImmutableHTableDescriptor.java
0
0
4
-
+
org/apache/hadoop/hbase/client/MasterCallable.java
0
0
-3
-
+2
+
org/apache/hadoop/hbase/client/MasterCoprocessorRpcChannelImpl.java
0
0
2
-
+
org/apache/hadoop/hbase/client/MasterKeepAliveConnection.java
0
0
1
-
+
org/apache/hadoop/hbase/client/MetaCache.java
0
0
6
-
+
org/apache/hadoop/hbase/client/MetricsConnection.java
0
0
41
-
+
org/apache/hadoop/hbase/client/MultiAction.java
0
0
3
-
+
org/apache/hadoop/hbase/client/MultiResponse.java
0
0
2
-
+
org/apache/hadoop/hbase/client/MultiServerCallable.java
0
0
6
-
+
org/apache/hadoop/hbase/client/Mutation.java
0
0
23
-
+
org/apache/hadoop/hbase/client/NoOpRetryableCallerInterceptor.java
0
0
1
-
+
org/apache/hadoop/hbase/client/NoncedRegionServerCallable.java
0
0
1
-
+
org/apache/hadoop/hbase/client/Operation.java
0
0
2
-
+
org/apache/hadoop/hbase/client/PackagePrivateFieldAccessor.java
0
0
1
-
+
org/apache/hadoop/hbase/client/PreemptiveFastFailInterceptor.java
0
0
13
-
+
org/apache/hadoop/hbase/client/Put.java
0
0
9
-
+
org/apache/hadoop/hbase/client/Query.java
0
0
10
-
+
org/apache/hadoop/hbase/client/QuotaStatusCalls.java
0
0
5
-
+
org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.java
0
0
88
-
+
org/apache/hadoop/hbase/client/RegionAdminServiceCallable.java
0
0
5
-
+
org/apache/hadoop/hbase/client/RegionCoprocessorRpcChannel.java
0
0
3
-
+
org/apache/hadoop/hbase/client/RegionCoprocessorRpcChannelImpl.java
0
0
1
-
+
org/apache/hadoop/hbase/client/RegionCoprocessorServiceExec.java
0
0
2
-
+
org/apache/hadoop/hbase/client/RegionInfo.java
0
0
55
-
+
org/apache/hadoop/hbase/client/RegionInfoBuilder.java
0
0
6
-
+
org/apache/hadoop/hbase/client/RegionInfoDisplay.java
0
0
17
-
+
org/apache/hadoop/hbase/client/RegionLocator.java
0
0
1
-
+
org/apache/hadoop/hbase/client/RegionReplicaUtil.java
0
0
8
-
+
org/apache/hadoop/hbase/client/RegionServerCallable.java
0
0
4
-
+
org/apache/hadoop/hbase/client/RegionServerCoprocessorRpcChannelImpl.java
0
0
2
-
+
org/apache/hadoop/hbase/client/RequestController.java
0
0
4
-
+
org/apache/hadoop/hbase/client/RequestControllerFactory.java
0
0
3
-
+
org/apache/hadoop/hbase/client/Result.java
0
0
13
-
+
org/apache/hadoop/hbase/client/ResultBoundedCompletionService.java
0
0
7
-
+
org/apache/hadoop/hbase/client/ResultScanner.java
0
0
3
-
+
org/apache/hadoop/hbase/client/RetriesExhaustedException.java
0
0
1
-
+
org/apache/hadoop/hbase/client/RetriesExhaustedWithDetailsException.java
0
0
3
-
+
org/apache/hadoop/hbase/client/RetryingCallerInterceptor.java
0
0
4
-
+
org/apache/hadoop/hbase/client/ReversedClientScanner.java
0
0
8
-
+
org/apache/hadoop/hbase/client/ReversedScannerCallable.java
0
0
11
-
+
org/apache/hadoop/hbase/client/RowAccess.java
0
0
1
-
+
org/apache/hadoop/hbase/client/RowMutations.java
0
0
1
-
+
org/apache/hadoop/hbase/client/RpcRetryingCaller.java
0
0
1
-
+
org/apache/hadoop/hbase/client/RpcRetryingCallerImpl.java
0
0
2
-
+
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/991224b9/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html
index e4e6219..6f4eff2 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html
@@ -18,7 +18,7 @@
catch(err) {
}
//-->
-var methods =
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":9,"i43":9,"i44":9,"i45":9,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":9,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":9,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":9,"i106":10,"i107":10,"i108":10,"i109":10
,"i110":10,"i111":10,"i112":41,"i113":41,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":9,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":10,"i145":42,"i146":10,"i147":10,"i148":10,"i149":10,"i150":10,"i151":10,"i152":10,"i153":10,"i154":10,"i155":10,"i156":10,"i157":10,"i158":10,"i159":10,"i160":10,"i161":10,"i162":10,"i163":10,"i164":10,"i165":10,"i166":10,"i167":9,"i168":10,"i169":10,"i170":10,"i171":10,"i172":10,"i173":10,"i174":10,"i175":10,"i176":9,"i177":10,"i178":10,"i179":9,"i180":9,"i181":9,"i182":9,"i183":9,"i184":9,"i185":9,"i186":9,"i187":9,"i188":9,"i189":10,"i190":10,"i191":10,"i192":10,"i193":10,"i194":10,"i195":10,"i196":10,"i197":10,"i198":9,"i199":10,"i200":10,"i201":10,"i202":10,"i203":10,"i204":10,"i205":10,"i206":10,"i207":10,"i208":10,"i209":10,"i210":10,
"i211":10,"i212":10,"i213":10,"i214":10,"i215":10,"i216":10,"i217":10,"i218":10,"i219":10,"i220":10,"i221":10,"i222":10,"i223":10,"i224":10,"i225":10,"i226":10,"i227":10,"i228":10,"i229":9,"i230":9,"i231":10,"i232":10,"i233":10,"i234":10,"i235":10,"i236":10,"i237":10,"i238":10,"i239":10,"i240":10,"i241":10,"i242":9,"i243":10,"i244":10,"i245":10,"i246":10,"i247":10,"i248":10,"i249":10,"i250":10,"i251":10,"i252":10,"i253":10,"i254":10,"i255":10,"i256":9,"i257":10,"i258":10,"i259":10,"i260":10};
+var methods =
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":9,"i42":9,"i43":9,"i44":9,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":9,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":9,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":9,"i106":10,"i107":10,"i108":10,"i109":10
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/org/apache/hadoop/hbase/master/class-use/HMaster.MasterStoppedException.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/class-use/HMaster.MasterStoppedException.html
b/devapidocs/org/apache/hadoop/hbase/master/class-use/HMaster.MasterStoppedException.html
new file mode 100644
index 000..985459e
--- /dev/null
+++
b/devapidocs/org/apache/hadoop/hbase/master/class-use/HMaster.MasterStoppedException.html
@@ -0,0 +1,165 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+
+Uses of Class
org.apache.hadoop.hbase.master.HMaster.MasterStoppedException (Apache HBase
3.0.0-SNAPSHOT API)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev
+Next
+
+
+Frames
+NoFrames
+
+
+AllClasses
+
+
+
+
+
+
+
+
+
+
+Uses of
Classorg.apache.hadoop.hbase.master.HMaster.MasterStoppedException
+
+
+
+
+
+Packages that use HMaster.MasterStoppedException
+
+Package
+Description
+
+
+
+org.apache.hadoop.hbase.master
+
+
+
+
+
+
+
+
+
+
+Uses of HMaster.MasterStoppedException in org.apache.hadoop.hbase.master
+
+Methods in org.apache.hadoop.hbase.master
that throw HMaster.MasterStoppedException
+
+Modifier and Type
+Method and Description
+
+
+
+(package private) void
+HMaster.checkInitialized()
+
+
+
+
+
+
+
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev
+Next
+
+
+Frames
+NoFrames
+
+
+AllClasses
+
+
+
+
+
+
+
+
+
+Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
+
+
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/org/apache/hadoop/hbase/master/package-frame.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/master/package-frame.html
b/devapidocs/org/apache/hadoop/hbase/master/package-frame.html
index b9dd127..7c5626e 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/package-frame.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/package-frame.html
@@ -106,6 +106,7 @@
Exceptions
ClusterSchemaException
+HMaster.MasterStoppedException
NoSuchProcedureException
TableStateManager.TableStateNotFoundException
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/org/apache/hadoop/hbase/master/package-summary.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/master/package-summary.html
b/devapidocs/org/apache/hadoop/hbase/master/package-summary.html
index 01cc345..81462ad 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/package-summary.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/package-summary.html
@@ -587,10 +587,14 @@
-NoSuchProcedureException
+HMaster.MasterStoppedException
+NoSuchProcedureException
+
+
+
TableStateManager.TableStateNotFoundException
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
index 6f6a834..b14f18f 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
@@ -240,6 +240,7 @@
org.apache.hadoop.hbase.master.ClusterSchemaException
org.apache.hadoop.hbase.DoNotRetryIOException
+org.apache.hadoop.hbase.master.HMaster.MasterStoppedException
org.apache.hadoop.hbase.TableNotFoundException
org.apache.hadoop.hbase.master.TableStateManager.TableStateNotFoundException
@@ -345,10 +346,10 @@
java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableT,
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/94208cfe/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
b/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
index 802d439..240226a 100644
--- a/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
+++ b/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
@@ -535,11 +535,6 @@ service.
static Cell
-PrivateCellUtil.createCell(Cellcell,
- byte[]tags)
-
-
-static Cell
CellUtil.createCell(Cellcell,
byte[]tags)
Deprecated.
@@ -547,13 +542,12 @@ service.
-
+
static Cell
-PrivateCellUtil.createCell(Cellcell,
- byte[]value,
+PrivateCellUtil.createCell(Cellcell,
byte[]tags)
-
+
static Cell
CellUtil.createCell(Cellcell,
byte[]value,
@@ -563,12 +557,13 @@ service.
-
+
static Cell
-PrivateCellUtil.createCell(Cellcell,
- http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListTagtags)
+PrivateCellUtil.createCell(Cellcell,
+ byte[]value,
+ byte[]tags)
-
+
static Cell
CellUtil.createCell(Cellcell,
http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListTagtags)
@@ -577,6 +572,11 @@ service.
+
+static Cell
+PrivateCellUtil.createCell(Cellcell,
+ http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListTagtags)
+
static Cell
PrivateCellUtil.createFirstDeleteFamilyCellOnRow(byte[]row,
@@ -757,10 +757,6 @@ service.
static byte[]
-PrivateCellUtil.cloneTags(Cellcell)
-
-
-static byte[]
CellUtil.cloneTags(Cellcell)
Deprecated.
As of HBase-2.0. Will be
removed in HBase-3.0.
@@ -768,6 +764,10 @@ service.
+
+static byte[]
+PrivateCellUtil.cloneTags(Cellcell)
+
static byte[]
CellUtil.cloneValue(Cellcell)
@@ -781,11 +781,6 @@ service.
int
-CellComparatorImpl.compare(Cella,
- Cellb)
-
-
-int
KeyValue.MetaComparator.compare(Cellleft,
Cellright)
Deprecated.
@@ -793,7 +788,7 @@ service.
table.
-
+
int
KeyValue.KVComparator.compare(Cellleft,
Cellright)
@@ -802,6 +797,11 @@ service.
rowkey, colfam/qual, timestamp, type, mvcc
+
+int
+CellComparatorImpl.compare(Cella,
+ Cellb)
+
int
CellComparatorImpl.compare(Cella,
@@ -812,27 +812,27 @@ service.
static int
-PrivateCellUtil.compare(CellComparatorcomparator,
+CellUtil.compare(CellComparatorcomparator,
Cellleft,
byte[]key,
intoffset,
intlength)
-Used when a cell needs to be compared with a key byte[]
such as cases of finding the index from
- the index block, bloom keys from the bloom blocks This byte[] is expected to
be serialized in
- the KeyValue serialization format If the KeyValue (Cell's) serialization
format changes this
- method cannot be used.
+Deprecated.
+As of HBase-2.0. Will be
removed in HBase-3.0
+
static int
-CellUtil.compare(CellComparatorcomparator,
+PrivateCellUtil.compare(CellComparatorcomparator,
Cellleft,
byte[]key,
intoffset,
intlength)
-Deprecated.
-As of HBase-2.0. Will be
removed in HBase-3.0
-
+Used when a cell needs to be compared with a key byte[]
such as cases of finding the index from
+ the index block, bloom keys from the bloom blocks This byte[] is expected to
be serialized in
+ the KeyValue serialization format If the KeyValue (Cell's) serialization
format changes this
+ method cannot be used.
@@ -1035,23 +1035,23 @@ service.
int
+KeyValue.KVComparator.compareRows(Cellleft,
+ Cellright)
+Deprecated.
+
+
+
+int
CellComparatorImpl.compareRows(Cellleft,
Cellright)
Compares the rows of the left and right cell.
-
+
int
CellComparatorImpl.MetaCellComparator.compareRows(Cellleft,
Cellright)
-
-int
-KeyValue.KVComparator.compareRows(Cellleft,
- Cellright)
-Deprecated.
-
-
int
CellComparator.compareTimestamps(CellleftCell,
@@ -1061,17 +1061,17 @@ service.
int
-CellComparatorImpl.compareTimestamps(Cellleft,
+KeyValue.KVComparator.compareTimestamps(Cellleft,
Cellright)
-Compares cell's timestamps in DESCENDING order.
-
+Deprecated.
+
int
-KeyValue.KVComparator.compareTimestamps(Cellleft,
+CellComparatorImpl.compareTimestamps(Cellleft,
Cellright)
-Deprecated.
-
+Compares cell's timestamps in DESCENDING order.
+
static int
@@ -1258,11 +1258,6 @@ service.
static Cell
-PrivateCellUtil.createCell(Cellcell,
- byte[]tags)
-
-
-static Cell
CellUtil.createCell(Cellcell,
byte[]tags)
Deprecated.
@@ -1270,13 +1265,12 @@ service.
-
+
static Cell
-PrivateCellUtil.createCell(Cellcell,
- byte[]value,
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cd17dc5/checkstyle.rss
--
diff --git a/checkstyle.rss b/checkstyle.rss
index 780d2d7..626a391 100644
--- a/checkstyle.rss
+++ b/checkstyle.rss
@@ -25,8 +25,8 @@ under the License.
en-us
2007 - 2018 The Apache Software Foundation
- File: 3534,
- Errors: 16577,
+ File: 3540,
+ Errors: 16555,
Warnings: 0,
Infos: 0
@@ -1894,7 +1894,7 @@ under the License.
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.mob.MobUtils.java;>org/apache/hadoop/hbase/mob/MobUtils.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.DateTieredStoreEngine.java;>org/apache/hadoop/hbase/regionserver/DateTieredStoreEngine.java
0
@@ -1903,12 +1903,12 @@ under the License.
0
- 15
+ 1
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.DateTieredStoreEngine.java;>org/apache/hadoop/hbase/regionserver/DateTieredStoreEngine.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.mob.MobUtils.java;>org/apache/hadoop/hbase/mob/MobUtils.java
0
@@ -1917,7 +1917,7 @@ under the License.
0
- 1
+ 15
@@ -3770,7 +3770,7 @@ under the License.
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.quotas.TestQuotaObserverChore.java;>org/apache/hadoop/hbase/quotas/TestQuotaObserverChore.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.util.HFileArchiveTestingUtil.java;>org/apache/hadoop/hbase/util/HFileArchiveTestingUtil.java
0
@@ -3779,12 +3779,12 @@ under the License.
0
- 0
+ 9
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.util.HFileArchiveTestingUtil.java;>org/apache/hadoop/hbase/util/HFileArchiveTestingUtil.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.quotas.TestQuotaObserverChore.java;>org/apache/hadoop/hbase/quotas/TestQuotaObserverChore.java
0
@@ -3793,7 +3793,7 @@ under the License.
0
- 9
+ 0
@@ -5338,7 +5338,7 @@ under the License.
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.TestRegionSnapshotTask.java;>org/apache/hadoop/hbase/snapshot/TestRegionSnapshotTask.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.TestSwitchToStreamRead.java;>org/apache/hadoop/hbase/regionserver/TestSwitchToStreamRead.java
0
@@ -5352,7 +5352,7 @@ under the License.
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.TestRestoreSnapshotHelper.java;>org/apache/hadoop/hbase/snapshot/TestRestoreSnapshotHelper.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.TestRegionSnapshotTask.java;>org/apache/hadoop/hbase/snapshot/TestRegionSnapshotTask.java
0
@@ -5361,12 +5361,12 @@ under the License.
0
- 2
+ 0
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.TestSwitchToStreamRead.java;>org/apache/hadoop/hbase/regionserver/TestSwitchToStreamRead.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.TestRestoreSnapshotHelper.java;>org/apache/hadoop/hbase/snapshot/TestRestoreSnapshotHelper.java
0
@@ -5375,7 +5375,7 @@ under the License.
0
-
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/checkstyle.rss
--
diff --git a/checkstyle.rss b/checkstyle.rss
index 361fd1b..780d2d7 100644
--- a/checkstyle.rss
+++ b/checkstyle.rss
@@ -25,8 +25,8 @@ under the License.
en-us
2007 - 2018 The Apache Software Foundation
- File: 3533,
- Errors: 16576,
+ File: 3534,
+ Errors: 16577,
Warnings: 0,
Infos: 0
@@ -1894,7 +1894,7 @@ under the License.
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.DateTieredStoreEngine.java;>org/apache/hadoop/hbase/regionserver/DateTieredStoreEngine.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.mob.MobUtils.java;>org/apache/hadoop/hbase/mob/MobUtils.java
0
@@ -1903,12 +1903,12 @@ under the License.
0
- 1
+ 15
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.mob.MobUtils.java;>org/apache/hadoop/hbase/mob/MobUtils.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.DateTieredStoreEngine.java;>org/apache/hadoop/hbase/regionserver/DateTieredStoreEngine.java
0
@@ -1917,7 +1917,7 @@ under the License.
0
- 15
+ 1
@@ -3770,7 +3770,7 @@ under the License.
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.util.HFileArchiveTestingUtil.java;>org/apache/hadoop/hbase/util/HFileArchiveTestingUtil.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.quotas.TestQuotaObserverChore.java;>org/apache/hadoop/hbase/quotas/TestQuotaObserverChore.java
0
@@ -3779,12 +3779,12 @@ under the License.
0
- 9
+ 0
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.quotas.TestQuotaObserverChore.java;>org/apache/hadoop/hbase/quotas/TestQuotaObserverChore.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.util.HFileArchiveTestingUtil.java;>org/apache/hadoop/hbase/util/HFileArchiveTestingUtil.java
0
@@ -3793,7 +3793,7 @@ under the License.
0
- 0
+ 9
@@ -5338,7 +5338,7 @@ under the License.
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.TestSwitchToStreamRead.java;>org/apache/hadoop/hbase/regionserver/TestSwitchToStreamRead.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.TestRegionSnapshotTask.java;>org/apache/hadoop/hbase/snapshot/TestRegionSnapshotTask.java
0
@@ -5352,7 +5352,7 @@ under the License.
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.TestRegionSnapshotTask.java;>org/apache/hadoop/hbase/snapshot/TestRegionSnapshotTask.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.TestRestoreSnapshotHelper.java;>org/apache/hadoop/hbase/snapshot/TestRestoreSnapshotHelper.java
0
@@ -5361,12 +5361,12 @@ under the License.
0
- 0
+ 2
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.TestRestoreSnapshotHelper.java;>org/apache/hadoop/hbase/snapshot/TestRestoreSnapshotHelper.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.TestSwitchToStreamRead.java;>org/apache/hadoop/hbase/regionserver/TestSwitchToStreamRead.java
0
@@ -5375,7 +5375,7 @@ under the License.
0
-
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/apidocs/overview-tree.html
--
diff --git a/apidocs/overview-tree.html b/apidocs/overview-tree.html
index 32feba1..d21d0ac 100644
--- a/apidocs/overview-tree.html
+++ b/apidocs/overview-tree.html
@@ -889,33 +889,33 @@
java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
title="class or interface in java.io">Serializable)
org.apache.hadoop.hbase.util.Order
-org.apache.hadoop.hbase.MemoryCompactionPolicy
org.apache.hadoop.hbase.KeepDeletedCells
+org.apache.hadoop.hbase.MemoryCompactionPolicy
org.apache.hadoop.hbase.CompareOperator
org.apache.hadoop.hbase.ProcedureState
org.apache.hadoop.hbase.CellBuilderType
+org.apache.hadoop.hbase.filter.BitComparator.BitwiseOp
org.apache.hadoop.hbase.filter.FilterList.Operator
org.apache.hadoop.hbase.filter.CompareFilter.CompareOp
-org.apache.hadoop.hbase.filter.BitComparator.BitwiseOp
-org.apache.hadoop.hbase.filter.RegexStringComparator.EngineType
org.apache.hadoop.hbase.filter.Filter.ReturnCode
+org.apache.hadoop.hbase.filter.RegexStringComparator.EngineType
org.apache.hadoop.hbase.io.encoding.DataBlockEncoding
org.apache.hadoop.hbase.regionserver.BloomType
-org.apache.hadoop.hbase.quotas.SpaceViolationPolicy
org.apache.hadoop.hbase.quotas.ThrottlingException.Type
org.apache.hadoop.hbase.quotas.QuotaScope
-org.apache.hadoop.hbase.quotas.QuotaType
org.apache.hadoop.hbase.quotas.ThrottleType
-org.apache.hadoop.hbase.client.SnapshotType
+org.apache.hadoop.hbase.quotas.QuotaType
+org.apache.hadoop.hbase.quotas.SpaceViolationPolicy
org.apache.hadoop.hbase.client.Durability
+org.apache.hadoop.hbase.client.SnapshotType
+org.apache.hadoop.hbase.client.MasterSwitchType
+org.apache.hadoop.hbase.client.CompactType
org.apache.hadoop.hbase.client.MobCompactPartitionPolicy
-org.apache.hadoop.hbase.client.IsolationLevel
-org.apache.hadoop.hbase.client.RequestController.ReturnCode
-org.apache.hadoop.hbase.client.Scan.ReadType
org.apache.hadoop.hbase.client.CompactionState
-org.apache.hadoop.hbase.client.MasterSwitchType
+org.apache.hadoop.hbase.client.Scan.ReadType
+org.apache.hadoop.hbase.client.RequestController.ReturnCode
+org.apache.hadoop.hbase.client.IsolationLevel
org.apache.hadoop.hbase.client.Consistency
-org.apache.hadoop.hbase.client.CompactType
org.apache.hadoop.hbase.client.security.SecurityCapability
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/book.html
--
diff --git a/book.html b/book.html
index 399814c..96d4058 100644
--- a/book.html
+++ b/book.html
@@ -30927,15 +30927,20 @@ For example, to skip the tests in
hbase-server and hbase-comm
167.2.
Unit Tests
-Apache HBase test cases are subdivided into four categories: small, medium,
large, and
-integration with corresponding JUnit https://github.com/junit-team/junit4/wiki/Categories;>categories:
SmallTests, MediumTests, LargeTests,
IntegrationTests.
-JUnit categories are denoted using java annotations and look like this in your
unit test code.
+Apache HBase unit tests must carry a Category annotation and
+as of hbase-2.0.0, must be stamped with the HBase
ClassRule.
+Here is an example of what a Test Class looks like with a
+Category and ClassRule included:
...
@Category(SmallTests.class)
public class TestHRegionInfo {
+ @ClassRule
+ public static
final HBaseClassTestRule CLASS_RULE =
+ HBaseClassTestRule.forClass(TestHRegionInfo.class);
+
@Test
public void
testCreateHRegionInfoName() throws Exception {
// ...
@@ -30944,8 +30949,23 @@ JUnit categories are denoted using java annotations
and look like this in your u
-The above example shows how to mark a test case as belonging to the
small category.
-All test cases in HBase should have a categorization.
+Here the Test Class is TestHRegionInfo. The
CLASS_RULE has
+the same form in every test class only the .class you pass
+is that of the local test; i.e. in the TestTimeout Test Class, youd
+pass TestTimeout.class to the CLASS_RULE instead of
the
+TestHRegionInfo.class we have above. The CLASS_RULE
+is where well enforce timeouts (currently set at a hard-limit of
+ten minutes for all tests) and other cross-unit test facility.
+The test is in the SmallTest Category.
+
+
+Categories can be arbitrary and provided as a list but each test MUST
+carry one from the following list of sizings: small,
medium, large, and
+integration. The test sizing is designated using the JUnit
+https://github.com/junit-team/junit4/wiki/Categories;>categories:
SmallTests,
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/checkstyle.rss
--
diff --git a/checkstyle.rss b/checkstyle.rss
index b6fad27..d597ba1 100644
--- a/checkstyle.rss
+++ b/checkstyle.rss
@@ -26,7 +26,7 @@ under the License.
2007 - 2018 The Apache Software Foundation
File: 3530,
- Errors: 16560,
+ Errors: 16557,
Warnings: 0,
Infos: 0
@@ -1894,7 +1894,7 @@ under the License.
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.DateTieredStoreEngine.java;>org/apache/hadoop/hbase/regionserver/DateTieredStoreEngine.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.mob.MobUtils.java;>org/apache/hadoop/hbase/mob/MobUtils.java
0
@@ -1903,12 +1903,12 @@ under the License.
0
- 1
+ 15
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.mob.MobUtils.java;>org/apache/hadoop/hbase/mob/MobUtils.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.DateTieredStoreEngine.java;>org/apache/hadoop/hbase/regionserver/DateTieredStoreEngine.java
0
@@ -1917,7 +1917,7 @@ under the License.
0
- 15
+ 1
@@ -3770,7 +3770,7 @@ under the License.
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.util.HFileArchiveTestingUtil.java;>org/apache/hadoop/hbase/util/HFileArchiveTestingUtil.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.quotas.TestQuotaObserverChore.java;>org/apache/hadoop/hbase/quotas/TestQuotaObserverChore.java
0
@@ -3779,12 +3779,12 @@ under the License.
0
- 9
+ 0
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.quotas.TestQuotaObserverChore.java;>org/apache/hadoop/hbase/quotas/TestQuotaObserverChore.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.util.HFileArchiveTestingUtil.java;>org/apache/hadoop/hbase/util/HFileArchiveTestingUtil.java
0
@@ -3793,7 +3793,7 @@ under the License.
0
- 0
+ 9
@@ -5338,7 +5338,7 @@ under the License.
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.TestSwitchToStreamRead.java;>org/apache/hadoop/hbase/regionserver/TestSwitchToStreamRead.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.TestRegionSnapshotTask.java;>org/apache/hadoop/hbase/snapshot/TestRegionSnapshotTask.java
0
@@ -5352,7 +5352,7 @@ under the License.
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.TestRegionSnapshotTask.java;>org/apache/hadoop/hbase/snapshot/TestRegionSnapshotTask.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.TestRestoreSnapshotHelper.java;>org/apache/hadoop/hbase/snapshot/TestRestoreSnapshotHelper.java
0
@@ -5361,12 +5361,12 @@ under the License.
0
- 0
+ 2
- http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.snapshot.TestRestoreSnapshotHelper.java;>org/apache/hadoop/hbase/snapshot/TestRestoreSnapshotHelper.java
+ http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.regionserver.TestSwitchToStreamRead.java;>org/apache/hadoop/hbase/regionserver/TestSwitchToStreamRead.java
0
@@ -5375,7 +5375,7 @@ under the License.
0
- 2
+
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0ab8335e/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
b/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
index 11ca943..e63037e 100644
--- a/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
+++ b/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
@@ -535,6 +535,11 @@ service.
static Cell
+PrivateCellUtil.createCell(Cellcell,
+ byte[]tags)
+
+
+static Cell
CellUtil.createCell(Cellcell,
byte[]tags)
Deprecated.
@@ -542,12 +547,13 @@ service.
-
+
static Cell
-PrivateCellUtil.createCell(Cellcell,
+PrivateCellUtil.createCell(Cellcell,
+ byte[]value,
byte[]tags)
-
+
static Cell
CellUtil.createCell(Cellcell,
byte[]value,
@@ -557,13 +563,12 @@ service.
-
+
static Cell
-PrivateCellUtil.createCell(Cellcell,
- byte[]value,
- byte[]tags)
+PrivateCellUtil.createCell(Cellcell,
+ http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListTagtags)
-
+
static Cell
CellUtil.createCell(Cellcell,
http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListTagtags)
@@ -572,11 +577,6 @@ service.
-
-static Cell
-PrivateCellUtil.createCell(Cellcell,
- http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListTagtags)
-
static Cell
PrivateCellUtil.createFirstDeleteFamilyCellOnRow(byte[]row,
@@ -757,6 +757,10 @@ service.
static byte[]
+PrivateCellUtil.cloneTags(Cellcell)
+
+
+static byte[]
CellUtil.cloneTags(Cellcell)
Deprecated.
As of HBase-2.0. Will be
removed in HBase-3.0.
@@ -764,10 +768,6 @@ service.
-
-static byte[]
-PrivateCellUtil.cloneTags(Cellcell)
-
static byte[]
CellUtil.cloneValue(Cellcell)
@@ -781,6 +781,11 @@ service.
int
+CellComparatorImpl.compare(Cella,
+ Cellb)
+
+
+int
KeyValue.MetaComparator.compare(Cellleft,
Cellright)
Deprecated.
@@ -788,7 +793,7 @@ service.
table.
-
+
int
KeyValue.KVComparator.compare(Cellleft,
Cellright)
@@ -797,11 +802,6 @@ service.
rowkey, colfam/qual, timestamp, type, mvcc
-
-int
-CellComparatorImpl.compare(Cella,
- Cellb)
-
int
CellComparatorImpl.compare(Cella,
@@ -812,27 +812,27 @@ service.
static int
-CellUtil.compare(CellComparatorcomparator,
+PrivateCellUtil.compare(CellComparatorcomparator,
Cellleft,
byte[]key,
intoffset,
intlength)
-Deprecated.
-As of HBase-2.0. Will be
removed in HBase-3.0
-
+Used when a cell needs to be compared with a key byte[]
such as cases of finding the index from
+ the index block, bloom keys from the bloom blocks This byte[] is expected to
be serialized in
+ the KeyValue serialization format If the KeyValue (Cell's) serialization
format changes this
+ method cannot be used.
static int
-PrivateCellUtil.compare(CellComparatorcomparator,
+CellUtil.compare(CellComparatorcomparator,
Cellleft,
byte[]key,
intoffset,
intlength)
-Used when a cell needs to be compared with a key byte[]
such as cases of finding the index from
- the index block, bloom keys from the bloom blocks This byte[] is expected to
be serialized in
- the KeyValue serialization format If the KeyValue (Cell's) serialization
format changes this
- method cannot be used.
+Deprecated.
+As of HBase-2.0. Will be
removed in HBase-3.0
+
@@ -1035,23 +1035,23 @@ service.
int
-KeyValue.KVComparator.compareRows(Cellleft,
- Cellright)
-Deprecated.
-
-
-
-int
CellComparatorImpl.compareRows(Cellleft,
Cellright)
Compares the rows of the left and right cell.
-
+
int
CellComparatorImpl.MetaCellComparator.compareRows(Cellleft,
Cellright)
+
+int
+KeyValue.KVComparator.compareRows(Cellleft,
+ Cellright)
+Deprecated.
+
+
int
CellComparator.compareTimestamps(CellleftCell,
@@ -1061,17 +1061,17 @@ service.
int
-KeyValue.KVComparator.compareTimestamps(Cellleft,
+CellComparatorImpl.compareTimestamps(Cellleft,
Cellright)
-Deprecated.
-
+Compares cell's timestamps in DESCENDING order.
+
int
-CellComparatorImpl.compareTimestamps(Cellleft,
+KeyValue.KVComparator.compareTimestamps(Cellleft,
Cellright)
-Compares cell's timestamps in DESCENDING order.
-
+Deprecated.
+
static int
@@ -1258,6 +1258,11 @@ service.
static Cell
+PrivateCellUtil.createCell(Cellcell,
+ byte[]tags)
+
+
+static Cell
CellUtil.createCell(Cellcell,
byte[]tags)
Deprecated.
@@ -1265,12 +1270,13 @@ service.
-
+
static Cell
-PrivateCellUtil.createCell(Cellcell,
+PrivateCellUtil.createCell(Cellcell,
+
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/org/apache/hadoop/hbase/master/SplitLogManager.ResubmitDirective.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/SplitLogManager.ResubmitDirective.html
b/devapidocs/org/apache/hadoop/hbase/master/SplitLogManager.ResubmitDirective.html
index 6e416db..225ac26 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/SplitLogManager.ResubmitDirective.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/SplitLogManager.ResubmitDirective.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
-public static enum SplitLogManager.ResubmitDirective
+public static enum SplitLogManager.ResubmitDirective
extends http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
title="class or interface in java.lang">EnumSplitLogManager.ResubmitDirective
@@ -210,7 +210,7 @@ the order they are declared.
CHECK
-public static finalSplitLogManager.ResubmitDirective CHECK
+public static finalSplitLogManager.ResubmitDirective CHECK
@@ -219,7 +219,7 @@ the order they are declared.
FORCE
-public static finalSplitLogManager.ResubmitDirective FORCE
+public static finalSplitLogManager.ResubmitDirective FORCE
@@ -236,7 +236,7 @@ the order they are declared.
values
-public staticSplitLogManager.ResubmitDirective[]values()
+public staticSplitLogManager.ResubmitDirective[]values()
Returns an array containing the constants of this enum
type, in
the order they are declared. This method may be used to iterate
over the constants as follows:
@@ -256,7 +256,7 @@ for (SplitLogManager.ResubmitDirective c :
SplitLogManager.ResubmitDirective.val
valueOf
-public staticSplitLogManager.ResubmitDirectivevalueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">Stringname)
+public staticSplitLogManager.ResubmitDirectivevalueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">Stringname)
Returns the enum constant of this type with the specified
name.
The string must match exactly an identifier used to declare an
enum constant in this type. (Extraneous whitespace characters are
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/org/apache/hadoop/hbase/master/SplitLogManager.Task.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/SplitLogManager.Task.html
b/devapidocs/org/apache/hadoop/hbase/master/SplitLogManager.Task.html
index c563b64..a3bb42d 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/SplitLogManager.Task.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/SplitLogManager.Task.html
@@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab";
@InterfaceAudience.Private
-public static class SplitLogManager.Task
+public static class SplitLogManager.Task
extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
in memory state of an active task.
@@ -253,7 +253,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
last_update
-public volatilelong last_update
+public volatilelong last_update
@@ -262,7 +262,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
last_version
-public volatileint last_version
+public volatileint last_version
@@ -271,7 +271,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
cur_worker_name
-public volatileServerName cur_worker_name
+public volatileServerName cur_worker_name
@@ -280,7 +280,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
batch
-public volatileSplitLogManager.TaskBatch
batch
+public volatileSplitLogManager.TaskBatch
batch
@@ -289,7 +289,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
status
-public volatileSplitLogManager.TerminationStatus status
+public volatileSplitLogManager.TerminationStatus status
@@ -298,7 +298,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
incarnation
-public volatilehttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicInteger.html?is-external=true;
title="class or interface in java.util.concurrent.atomic">AtomicInteger incarnation
+public volatilehttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicInteger.html?is-external=true;
title="class or interface in java.util.concurrent.atomic">AtomicInteger incarnation
@@ -307,7 +307,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
unforcedResubmits
-public finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicInteger.html?is-external=true;
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1f2eeb22/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html
b/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html
index 2fb2450..147ec8c 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/HMasterCommandLine.LocalHMaster.html
@@ -247,7 +247,7 @@ extends
Methods inherited from classorg.apache.hadoop.hbase.master.HMaster
-abort,
abortProcedure,
addColumn,
addReplicationPeer,
balance,
balance,
balanceSwitch,
canCreateBaseZNode, canUpdateTableDescriptor,
checkIfShouldMoveSystemRegionAsync,
checkInitialized,
checkServiceStarted,
checkTableModifiable,
configureInfoServer,
constructMaster,
createMetaBootstrap,
createNamespace,
createQuotaSnapshotNotifier,
createRpcServices,
createServerManager,
createSystemTable,
createTable,
decommissionRegionServers,
deleteColumn,
deleteNamespace,
deleteTable,
disableReplicationPeer,
disableTable,
enableReplicationPeer,
enableTable,
getAssignmentManager,
getAverageLoad,
getCatalogJanitor,
getClientIdAuditPrefix,
getClusterMetrics,
getClusterMetrics,
getClusterMetricsWithoutCoprocessor, getClusterMetricsWithoutCoprocessor,
getClusterSchema,
getDumpServlet,
getFavoredNodesManager,
getHFileCleaner,
getInitializedEvent,
getLastMajorCompactionTimestamp,
getLastMajorCompactionTimestampForRegion,
getLoadBalancer,
getLoadBalancerClassName,
getLoadedCoprocessors,
getLockManager,
getLocks,
getLogCleaner,
getMasterActiveTime,
getMasterCoprocessorHost, getMasterCoprocessors,
getMasterFileSystem,
getMasterFinishedInitializationTime,
getMasterMetrics,
getMasterProcedureExecutor,
getMasterProcedureManagerHost,
getMasterQuotaManager,
getMasterRpcServices, getMasterStartTime,
getMasterWalManager,
getMergePlanCount,
getMetaTableObserver,
getMobCompactionState,
getNamespace,
getNamespaces,
getNumWALFiles,
getProcedures,
getProcessName,
getQuotaObserverChore,
getRegionNormalizer,
getRegionNormalizerTracker,
getRegionServerFatalLogBuffer,
getRegionServerInfoPort,
getRegionSe
rverVersion, getRemoteInetAddress,
getReplicationPeerConfig,
getReplicationPeerManager,
getServerCrashProcessingEnabledEvent,
getServerManager,
getServerName,
getSnapshotManager,
getSp
aceQuotaSnapshotNotifier, getSplitOrMergeTracker,
getSplitPlanCount,
getTableDescriptors,
getTableRegionForRow,
getTableStateManager,
getUseThisHostnameInstead,
getWalProcedureStore,
getZooKeeper, initClusterSchemaService,
initializeZKBasedSystemTrackers,
initQuotaManager,
isActiveMaster,
isBalancerOn,
isCatalogJanitorEnabled,
isCleanerChoreEnabled,
isInitialized,
isInMaintenanceMode,
isNormalizerOn,
isServerCrashProcessingEnabled,
isSplitOrMergeEnabled,
listDecommissionedRegionServers,
listReplicationPeers,
listTableDescriptors,
listTableDescriptorsByNamespace,
listTableNames,
listTableNamesByNamespace,
login,
main,
mergeRegions,
modifyColumn, modifyNamespace,
modifyTable,
move,
normalizeRegions,
recommissionRegionServer,
recoverMeta,
registerService,
remoteProcedureCompleted,
remoteProcedureFailed,
removeReplicationPeer,
reportMobCompactionEnd,
reportMobCompactionStart,
requestMobCompaction,
restoreSnapshot,
setCatalogJanitorEnabled,
setInitialized,
setServerCrashProcessingEnabled,
shutdown,
splitRegion,
stop,
stopMaster,
stopServiceThreads,
truncateTable,
updateConfigurationForSpaceQuotaObserver,
updateReplicationPeerConfig,
waitForMasterActive
+abort,
abortProcedure,
addColumn,
addReplicationPeer,
balance,
balance,
balanceSwitch,
canCreateBaseZNode, canUpdateTableDescriptor,
checkIfShouldMoveSystemRegionAsync,
checkInitialized,
checkServiceStarted,
checkTableModifiable,
configureInfoServer,
constructMaster,
createMetaBootstrap,
createNamespace,
createQuotaSnapshotNotifier,
createRpcServices,
createServerManager,
createSystemTable,
createTable,
decommissionRegionServers,
decorateMasterConfiguration,
deleteColumn,
deleteNamespace,
deleteTable,
disableReplicationPeer,
disableTable,
enableReplicationPeer,
enableTable,
getAssignmentManager,
getAverageLoad,
getCatalogJanitor,
getClientIdAuditPrefix,
getClusterMetrics,
getClusterMetrics,
getClusterMetricsWithoutCoprocessor,
getClusterMetricsWithoutCoprocessor,
getClusterSchema,
getDumpServlet,
getFavoredNodesManager,
getHFileCleaner,
getInitializedEvent,
getLastMajorCompactionTimestamp,
getLastMajorCompactionTimestampForRegion,
getLoadBalancer,
getLoadBalancerClassName,
getLoadedCoprocessors,
getLockManager,
getLocks,
getLogCleaner,
getMasterActiveTime,
getMasterCoprocessorHost,
getMasterCoprocessors,
getMasterFileSystem,
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/hbase-build-configuration/license.html
--
diff --git a/hbase-build-configuration/license.html
b/hbase-build-configuration/license.html
index e60239b..398de86 100644
--- a/hbase-build-configuration/license.html
+++ b/hbase-build-configuration/license.html
@@ -7,7 +7,7 @@
-
+
Apache HBase - Build Configuration Project Licenses
@@ -326,7 +326,7 @@
https://www.apache.org/;>The Apache Software
Foundation.
All rights reserved.
- Last Published:
2018-01-28
+ Last Published:
2018-01-29
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/hbase-build-configuration/mail-lists.html
--
diff --git a/hbase-build-configuration/mail-lists.html
b/hbase-build-configuration/mail-lists.html
index 105aeff..059e207 100644
--- a/hbase-build-configuration/mail-lists.html
+++ b/hbase-build-configuration/mail-lists.html
@@ -7,7 +7,7 @@
-
+
Apache HBase - Build Configuration Project Mailing
Lists
@@ -176,7 +176,7 @@
https://www.apache.org/;>The Apache Software
Foundation.
All rights reserved.
- Last Published:
2018-01-28
+ Last Published:
2018-01-29
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/hbase-build-configuration/plugin-management.html
--
diff --git a/hbase-build-configuration/plugin-management.html
b/hbase-build-configuration/plugin-management.html
index 44d1f74..c87c784 100644
--- a/hbase-build-configuration/plugin-management.html
+++ b/hbase-build-configuration/plugin-management.html
@@ -7,7 +7,7 @@
-
+
Apache HBase - Build Configuration Project Plugin
Management
@@ -271,7 +271,7 @@
https://www.apache.org/;>The Apache Software
Foundation.
All rights reserved.
- Last Published:
2018-01-28
+ Last Published:
2018-01-29
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/hbase-build-configuration/plugins.html
--
diff --git a/hbase-build-configuration/plugins.html
b/hbase-build-configuration/plugins.html
index b5720d0..7d2e92c 100644
--- a/hbase-build-configuration/plugins.html
+++ b/hbase-build-configuration/plugins.html
@@ -7,7 +7,7 @@
-
+
Apache HBase - Build Configuration Project Plugins
@@ -214,7 +214,7 @@
https://www.apache.org/;>The Apache Software
Foundation.
All rights reserved.
- Last Published:
2018-01-28
+ Last Published:
2018-01-29
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/hbase-build-configuration/project-info.html
--
diff --git a/hbase-build-configuration/project-info.html
b/hbase-build-configuration/project-info.html
index 2333f6d..0bf6071 100644
--- a/hbase-build-configuration/project-info.html
+++ b/hbase-build-configuration/project-info.html
@@ -7,7 +7,7 @@
-
+
Apache HBase - Build Configuration Project
Information
@@ -167,7 +167,7 @@
https://www.apache.org/;>The Apache Software
Foundation.
All rights reserved.
- Last Published:
2018-01-28
+ Last Published:
2018-01-29
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/hbase-build-configuration/project-summary.html
--
diff --git a/hbase-build-configuration/project-summary.html
b/hbase-build-configuration/project-summary.html
index 204d645..4cf83e9 100644
--- a/hbase-build-configuration/project-summary.html
+++ b/hbase-build-configuration/project-summary.html
@@ -7,7 +7,7 @@
-
+
Apache HBase - Build Configuration Project Summary
@@ -163,7 +163,7 @@
https://www.apache.org/;>The Apache Software
Foundation.
All rights reserved.
- Last Published:
2018-01-28
+ Last Published:
2018-01-29
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa7ffc92/devapidocs/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html
b/devapidocs/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html
index fb75405..138369a 100644
--- a/devapidocs/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html
+++ b/devapidocs/org/apache/hadoop/hbase/backup/impl/FullTableBackupClient.html
@@ -202,7 +202,7 @@ extends
void
execute()
-Backup request execution
+Backup request execution.
@@ -319,14 +319,14 @@ extends
execute
-publicvoidexecute()
+publicvoidexecute()
throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
-Backup request execution
+Backup request execution.
Specified by:
executein
classTableBackupClient
Throws:
-http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
+http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException - if the
execution of the backup fails
@@ -336,7 +336,7 @@ extends
snapshotTable
-protectedvoidsnapshotTable(Adminadmin,
+protectedvoidsnapshotTable(Adminadmin,
TableNametableName,
http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">StringsnapshotName)
throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa7ffc92/devapidocs/org/apache/hadoop/hbase/backup/impl/IncrementalBackupManager.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/backup/impl/IncrementalBackupManager.html
b/devapidocs/org/apache/hadoop/hbase/backup/impl/IncrementalBackupManager.html
index c7279d7..a30f769 100644
---
a/devapidocs/org/apache/hadoop/hbase/backup/impl/IncrementalBackupManager.html
+++
b/devapidocs/org/apache/hadoop/hbase/backup/impl/IncrementalBackupManager.html
@@ -219,7 +219,7 @@ extends
http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
getIncrBackupLogFileList()
-Get list of WAL files eligible for incremental backup
+Get list of WAL files eligible for incremental backup.
@@ -346,12 +346,12 @@ extends http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">StringgetIncrBackupLogFileList()
throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
-Get list of WAL files eligible for incremental backup
+Get list of WAL files eligible for incremental backup.
Returns:
list of WAL files
Throws:
-http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
+http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException - if getting the
list of WAL files fails
@@ -361,7 +361,7 @@ extends
excludeAlreadyBackedUpWALs
-privatehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">StringexcludeAlreadyBackedUpWALs(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">StringlogList,
+privatehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">StringexcludeAlreadyBackedUpWALs(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">StringlogList,
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/96e5e102/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
index ca8be5e..b8e6dfa 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
@@ -6398,514 +6398,514 @@
6390 int initialBatchProgress =
scannerContext.getBatchProgress();
6391 long initialSizeProgress =
scannerContext.getDataSizeProgress();
6392 long initialHeapSizeProgress =
scannerContext.getHeapSizeProgress();
-6393 long initialTimeProgress =
scannerContext.getTimeProgress();
-6394
-6395 // The loop here is used only when
at some point during the next we determine
-6396 // that due to effects of filters
or otherwise, we have an empty row in the result.
-6397 // Then we loop and try again.
Otherwise, we must get out on the first iteration via return,
-6398 // "true" if there's more data to
read, "false" if there isn't (storeHeap is at a stop row,
-6399 // and joinedHeap has no more data
to read for the last row (if set, joinedContinuationRow).
-6400 while (true) {
-6401// Starting to scan a new row.
Reset the scanner progress according to whether or not
-6402// progress should be kept.
-6403if
(scannerContext.getKeepProgress()) {
-6404 // Progress should be kept.
Reset to initial values seen at start of method invocation.
-6405
scannerContext.setProgress(initialBatchProgress, initialSizeProgress,
-6406 initialHeapSizeProgress,
initialTimeProgress);
-6407} else {
-6408
scannerContext.clearProgress();
-6409}
-6410if (rpcCall.isPresent()) {
-6411 // If a user specifies a
too-restrictive or too-slow scanner, the
-6412 // client might time out and
disconnect while the server side
-6413 // is still processing the
request. We should abort aggressively
-6414 // in that case.
-6415 long afterTime =
rpcCall.get().disconnectSince();
-6416 if (afterTime = 0) {
-6417throw new
CallerDisconnectedException(
-6418"Aborting on region " +
getRegionInfo().getRegionNameAsString() + ", call " +
-6419this + " after " +
afterTime + " ms, since " +
-6420"caller
disconnected");
-6421 }
-6422}
-6423
-6424// Let's see what we have in the
storeHeap.
-6425Cell current =
this.storeHeap.peek();
-6426
-6427boolean shouldStop =
shouldStop(current);
-6428// When has filter row is true
it means that the all the cells for a particular row must be
-6429// read before a filtering
decision can be made. This means that filters where hasFilterRow
-6430// run the risk of
enLongAddering out of memory errors in the case that they are applied to a
-6431// table that has very large
rows.
-6432boolean hasFilterRow =
this.filter != null this.filter.hasFilterRow();
-6433
-6434// If filter#hasFilterRow is
true, partial results are not allowed since allowing them
-6435// would prevent the filters
from being evaluated. Thus, if it is true, change the
-6436// scope of any limits that
could potentially create partial results to
-6437// LimitScope.BETWEEN_ROWS so
that those limits are not reached mid-row
-6438if (hasFilterRow) {
-6439 if (LOG.isTraceEnabled()) {
-6440
LOG.trace("filter#hasFilterRow is true which prevents partial results from
being "
-6441+ " formed. Changing
scope of limits that may create partials");
-6442 }
-6443
scannerContext.setSizeLimitScope(LimitScope.BETWEEN_ROWS);
-6444
scannerContext.setTimeLimitScope(LimitScope.BETWEEN_ROWS);
-6445}
-6446
-6447// Check if we were getting data
from the joinedHeap and hit the limit.
-6448// If not, then it's main path -
getting results from storeHeap.
-6449if (joinedContinuationRow ==
null) {
-6450 // First, check if we are at a
stop row. If so, there are no more results.
-6451 if (shouldStop) {
-6452if (hasFilterRow) {
-6453
filter.filterRowCells(results);
-6454}
-6455return
scannerContext.setScannerState(NextState.NO_MORE_VALUES).hasMoreValues();
-6456 }
-6457
-6458 // Check if rowkey filter
wants to exclude this row. If so, loop to next.
-6459 // Technically, if we hit
limits before on this row, we don't need this call.
-6460 if
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8118541f/checkstyle.rss
--
diff --git a/checkstyle.rss b/checkstyle.rss
index 7814192..4b9c840 100644
--- a/checkstyle.rss
+++ b/checkstyle.rss
@@ -25,8 +25,8 @@ under the License.
en-us
2007 - 2018 The Apache Software Foundation
- File: 3500,
- Errors: 18095,
+ File: 3503,
+ Errors: 17563,
Warnings: 0,
Infos: 0
@@ -531,7 +531,7 @@ under the License.
0
- 11
+ 2
@@ -797,7 +797,7 @@ under the License.
0
- 7
+ 6
@@ -895,7 +895,7 @@ under the License.
0
- 8
+ 0
@@ -2687,7 +2687,7 @@ under the License.
0
- 7
+ 6
@@ -3275,7 +3275,7 @@ under the License.
0
- 10
+ 3
@@ -4591,7 +4591,7 @@ under the License.
0
- 10
+ 9
@@ -6383,7 +6383,7 @@ under the License.
0
- 8
+ 0
@@ -6635,7 +6635,7 @@ under the License.
0
- 25
+ 23
@@ -6691,7 +6691,7 @@ under the License.
0
- 3
+ 0
@@ -6789,7 +6789,7 @@ under the License.
0
- 201
+ 14
@@ -8581,7 +8581,7 @@ under the License.
0
- 52
+ 17
@@ -8609,7 +8609,7 @@ under the License.
0
- 7
+ 9
@@ -8805,7 +8805,7 @@ under the License.
0
- 1
+ 0
@@ -9561,7 +9561,7 @@ under the License.
0
- 7
+ 1
@@ -11227,7 +11227,7 @@ under the License.
0
- 31
+ 30
@@ -11325,7 +11325,7 @@ under the License.
0
- 32
+ 3
@@ -12711,7 +12711,7 @@ under the License.
0
- 1
+ 0
@@ -13033,7 +13033,7 @@ under the License.
0
- 1
+ 0
@@ -14069,7 +14069,7 @@ under the License.
0
- 6
+ 5
@@ -15357,7 +15357,7 @@ under the License.
0
- 2
+ 1
@@ -15637,7 +15637,7 @@ under the License.
0
- 2
+ 1
@@ -15875,7 +15875,7 @@ under the License.
0
- 13
+ 11
@@
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/14db89d7/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html
index b50a65f..7271567 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.CompletedProcedureCleaner.html
@@ -1718,312 +1718,314 @@
1710
1711public WorkerThread(final
ThreadGroup group) {
1712 super(group, "ProcExecWrkr-" +
workerId.incrementAndGet());
-1713}
-1714
-1715@Override
-1716public void sendStopSignal() {
-1717 scheduler.signalAll();
-1718}
-1719
-1720@Override
-1721public void run() {
-1722 long lastUpdate =
EnvironmentEdgeManager.currentTime();
-1723 try {
-1724while (isRunning()
keepAlive(lastUpdate)) {
-1725 this.activeProcedure =
scheduler.poll(keepAliveTime, TimeUnit.MILLISECONDS);
-1726 if (this.activeProcedure ==
null) continue;
-1727 int activeCount =
activeExecutorCount.incrementAndGet();
-1728 int runningCount =
store.setRunningProcedureCount(activeCount);
-1729 if (LOG.isTraceEnabled()) {
-1730LOG.trace("Execute pid=" +
this.activeProcedure.getProcId() +
-1731" runningCount=" +
runningCount + ", activeCount=" + activeCount);
-1732 }
-1733
executionStartTime.set(EnvironmentEdgeManager.currentTime());
-1734 try {
-1735
executeProcedure(this.activeProcedure);
-1736 } catch (AssertionError e) {
-1737LOG.info("ASSERT pid=" +
this.activeProcedure.getProcId(), e);
-1738throw e;
-1739 } finally {
-1740activeCount =
activeExecutorCount.decrementAndGet();
-1741runningCount =
store.setRunningProcedureCount(activeCount);
-1742if (LOG.isTraceEnabled())
{
-1743 LOG.trace("Halt pid=" +
this.activeProcedure.getProcId() +
-1744 " runningCount=" +
runningCount + ", activeCount=" + activeCount);
-1745}
-1746this.activeProcedure =
null;
-1747lastUpdate =
EnvironmentEdgeManager.currentTime();
-1748
executionStartTime.set(Long.MAX_VALUE);
-1749 }
-1750}
-1751 } catch (Throwable t) {
-1752LOG.warn("Worker terminating
UNNATURALLY " + this.activeProcedure, t);
-1753 } finally {
-1754LOG.debug("Worker
terminated.");
-1755 }
-1756 workerThreads.remove(this);
-1757}
-1758
-1759@Override
-1760public String toString() {
-1761 Procedure? p =
this.activeProcedure;
-1762 return getName() + "(pid=" + (p ==
null? Procedure.NO_PROC_ID: p.getProcId() + ")");
-1763}
-1764
-1765/**
-1766 * @return the time since the
current procedure is running
-1767 */
-1768public long getCurrentRunTime() {
-1769 return
EnvironmentEdgeManager.currentTime() - executionStartTime.get();
-1770}
-1771
-1772private boolean keepAlive(final long
lastUpdate) {
-1773 if (workerThreads.size() =
corePoolSize) return true;
-1774 return
(EnvironmentEdgeManager.currentTime() - lastUpdate) keepAliveTime;
-1775}
-1776 }
-1777
-1778 /**
-1779 * Runs task on a period such as check
for stuck workers.
-1780 * @see InlineChore
-1781 */
-1782 private final class
TimeoutExecutorThread extends StoppableThread {
-1783private final
DelayQueueDelayedWithTimeout queue = new DelayQueue();
-1784
-1785public TimeoutExecutorThread(final
ThreadGroup group) {
-1786 super(group, "ProcExecTimeout");
-1787}
-1788
-1789@Override
-1790public void sendStopSignal() {
-1791
queue.add(DelayedUtil.DELAYED_POISON);
-1792}
-1793
-1794@Override
-1795public void run() {
-1796 final boolean traceEnabled =
LOG.isTraceEnabled();
-1797 while (isRunning()) {
-1798final DelayedWithTimeout task =
DelayedUtil.takeWithoutInterrupt(queue);
-1799if (task == null || task ==
DelayedUtil.DELAYED_POISON) {
-1800 // the executor may be
shutting down,
-1801 // and the task is just the
shutdown request
-1802 continue;
-1803}
-1804
-1805if (traceEnabled) {
-1806 LOG.trace("Executing " +
task);
-1807}
-1808
-1809// execute the task
-1810if (task instanceof InlineChore)
{
-1811
execInlineChore((InlineChore)task);
-1812} else if (task instanceof
DelayedProcedure) {
-1813
execDelayedProcedure((DelayedProcedure)task);
-1814} else {
-1815
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/devapidocs/org/apache/hadoop/hbase/zookeeper/ZooKeeperHelper.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/zookeeper/ZooKeeperHelper.html
b/devapidocs/org/apache/hadoop/hbase/zookeeper/ZooKeeperHelper.html
new file mode 100644
index 000..6d0efdd
--- /dev/null
+++ b/devapidocs/org/apache/hadoop/hbase/zookeeper/ZooKeeperHelper.html
@@ -0,0 +1,314 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+
+ZooKeeperHelper (Apache HBase 3.0.0-SNAPSHOT API)
+
+
+
+
+
+var methods = {"i0":9,"i1":9};
+var tabs = {65535:["t0","All Methods"],1:["t1","Static
Methods"],8:["t4","Concrete Methods"]};
+var altColor = "altColor";
+var rowColor = "rowColor";
+var tableTab = "tableTab";
+var activeTableTab = "activeTableTab";
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+PrevClass
+NextClass
+
+
+Frames
+NoFrames
+
+
+AllClasses
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.hadoop.hbase.zookeeper
+Class ZooKeeperHelper
+
+
+
+http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">java.lang.Object
+
+
+org.apache.hadoop.hbase.zookeeper.ZooKeeperHelper
+
+
+
+
+
+
+
+
+@InterfaceAudience.Private
+public final class ZooKeeperHelper
+extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
+Methods that help working with ZooKeeper
+
+
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors
+
+Modifier
+Constructor and Description
+
+
+private
+ZooKeeperHelper()
+
+
+
+
+
+
+
+
+
+Method Summary
+
+All MethodsStatic MethodsConcrete Methods
+
+Modifier and Type
+Method and Description
+
+
+static org.apache.zookeeper.ZooKeeper
+ensureConnectedZooKeeper(org.apache.zookeeper.ZooKeeperzookeeper,
+inttimeout)
+Ensure passed zookeeper is connected.
+
+
+
+static org.apache.zookeeper.ZooKeeper
+getConnectedZooKeeper(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">StringconnectString,
+ intsessionTimeoutMs)
+Get a ZooKeeper instance and wait until it connected before
returning.
+
+
+
+
+
+
+
+Methods inherited from classjava.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
+http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--;
title="class or interface in java.lang">clone, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-;
title="class or interface in java.lang">equals, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--;
title="class or interface in java.lang">finalize, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--;
title="class or interface in java.lang">getClass, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--;
title="class or interface in java.lang">hashCode, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--;
title="class or interface in java.lang">notify, http://docs.oracle.com/javase/8/docs/api/java/lang
/Object.html?is-external=true#notifyAll--" title="class or interface in
java.lang">notifyAll, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--;
title="class or interface in java.lang">toString, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait--;
title="class or interface in java.lang">wait, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-;
title="class or interface in java.lang">wait, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-int-;
title="class or interface in java.lang">wait
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Constructor Detail
+
+
+
+
+
+ZooKeeperHelper
+privateZooKeeperHelper()
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+getConnectedZooKeeper
+public staticorg.apache.zookeeper.ZooKeepergetConnectedZooKeeper(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f183e80f/devapidocs/org/apache/hadoop/hbase/wal/RegionGroupingProvider.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/wal/RegionGroupingProvider.html
b/devapidocs/org/apache/hadoop/hbase/wal/RegionGroupingProvider.html
index 02ca11f..97a2ac5 100644
--- a/devapidocs/org/apache/hadoop/hbase/wal/RegionGroupingProvider.html
+++ b/devapidocs/org/apache/hadoop/hbase/wal/RegionGroupingProvider.html
@@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab";
@InterfaceAudience.Private
-public class RegionGroupingProvider
+public class RegionGroupingProvider
extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
implements WALProvider
A WAL Provider that returns a WAL per group of regions.
@@ -194,7 +194,7 @@ implements
-private IdLock
+private KeyLockerhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
createLock
@@ -307,8 +307,7 @@ implements
WAL
-getWAL(byte[]identifier,
- byte[]namespace)
+getWAL(RegionInforegion)
private WAL
@@ -361,7 +360,7 @@ implements
LOG
-private static finalorg.slf4j.Logger LOG
+private static finalorg.slf4j.Logger LOG
@@ -370,7 +369,7 @@ implements
REGION_GROUPING_STRATEGY
-public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String REGION_GROUPING_STRATEGY
+public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String REGION_GROUPING_STRATEGY
See Also:
Constant
Field Values
@@ -383,7 +382,7 @@ implements
DEFAULT_REGION_GROUPING_STRATEGY
-public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String DEFAULT_REGION_GROUPING_STRATEGY
+public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String DEFAULT_REGION_GROUPING_STRATEGY
@@ -392,7 +391,7 @@ implements
DELEGATE_PROVIDER
-public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String DELEGATE_PROVIDER
+public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String DELEGATE_PROVIDER
delegate provider for WAL creation/roll/close
See Also:
@@ -406,7 +405,7 @@ implements
DEFAULT_DELEGATE_PROVIDER
-public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String DEFAULT_DELEGATE_PROVIDER
+public static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String DEFAULT_DELEGATE_PROVIDER
@@ -415,7 +414,7 @@ implements
META_WAL_GROUP_NAME
-private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String META_WAL_GROUP_NAME
+private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String META_WAL_GROUP_NAME
See Also:
Constant
Field Values
@@ -428,7 +427,7 @@ implements
cached
-private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true;
title="class or interface in java.util.concurrent">ConcurrentMaphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String,WALProvider cached
+private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ConcurrentMap.html?is-external=true;
title="class or interface in java.util.concurrent">ConcurrentMaphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String,WALProvider cached
A group-provider mapping, make sure one-one rather than
many-one mapping
@@ -438,7 +437,7 @@ implements
createLock
-private finalIdLock createLock
+private finalKeyLockerhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String createLock
@@ -447,7 +446,7 @@ implements
strategy
-privateRegionGroupingProvider.RegionGroupingStrategy
strategy
+privateRegionGroupingProvider.RegionGroupingStrategy
strategy
@@ -456,7 +455,7 @@ implements
factory
-privateWALFactory factory
+privateWALFactory factory
@@ -465,7 +464,7 @@ implements
listeners
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/book.html
--
diff --git a/book.html b/book.html
index 6fd821f..e5d84fa 100644
--- a/book.html
+++ b/book.html
@@ -12111,6 +12111,10 @@ For an example of using both together, see Se
property
+ namehbase.security.authorization/name
+ valuetrue/value
+/property
+property
namehbase.coprocessor.region.classes/name
valueorg.apache.hadoop.hbase.security.access.AccessController,
org.apache.hadoop.hbase.security.token.TokenProvider/value
/property
@@ -12640,6 +12644,10 @@ These properties take a list of class names.
property
+ namehbase.security.authorization/name
+ valuetrue/value
+/property
+property
namehbase.coprocessor.region.classes/name
valueorg.apache.hadoop.hbase.security.visibility.VisibilityController/value
/property
@@ -12992,6 +13000,10 @@ public Void run()
t
property
+ namehbase.security.authorization/name
+ valuetrue/value
+/property
+property
namehbase.coprocessor.regionserver.classes/name
valueorg.apache.hadoop.hbase.security.visibility.VisibilityController$VisibilityReplication/value
/property
@@ -13292,6 +13304,10 @@ HBase manages creation and deletion of this
directory.
property
+ namehbase.security.authorization/name
+ valuetrue/value
+/property
+property
namehbase.bulkload.staging.dir/name
value/tmp/hbase-staging/value
/property
@@ -13303,6 +13319,24 @@ HBase manages creation and deletion of this
directory.
+
+63.6. Secure Enable
+
+After hbase-2.x, the default 'hbase.security.authorization' changed.
+Before hbase-2.x, it defaulted to true, in later HBase versions, the
+default became false.
+So to enable hbase authorization, the following propertie must be configured
in hbase-site.xml.
+See https://issues.apache.org/jira/browse/HBASE-19483;>HBASE-19483;
+
+
+
+property
+ namehbase.security.authorization/name
+ valuetrue/value
+/property
+
+
+
@@ -13329,6 +13363,10 @@ All options have been discussed separately in the
sections above.
/property
!-- Coprocessors for ACLs and Visibility Tags
--
property
+ namehbase.security.authorization/name
+ valuetrue/value
+/property
+property
namehbase.coprocessor.region.classes/name
valueorg.apache.hadoop.hbase.security.access.AccessController,
org.apache.hadoop.hbase.security.visibility.VisibilityController,
@@ -29461,6 +29499,20 @@ before you start moving out the dead. Move in good
live nodes first if you have
If it appears stuck, restart the Master process.
+
+157.3. ACL
+
+To enable ACL, add the following to your hbase-site.xml and restart your
Master:
+
+
+
+property
+ namehbase.security.authorization/name
+ valuetrue/value
+property
+
+
+
@@ -35297,6 +35349,61 @@ In case the table goes out of date, the unit tests
which check for accuracy of p
replicateLogEntries
superuser|global(W)
+
+RSGroup
+addRSGroup
+superuser|global(A)
+
+
+
+balanceRSGroup
+superuser|global(A)
+
+
+
+getRSGroupInfo
+superuser|global(A)
+
+
+
+getRSGroupInfoOfTable
+superuser|global(A)
+
+
+
+getRSGroupOfServer
+superuser|global(A)
+
+
+
+listRSGroups
+superuser|global(A)
+
+
+
+moveServers
+superuser|global(A)
+
+
+
+moveServersAndTables
+superuser|global(A)
+
+
+
+moveTables
+superuser|global(A)
+
+
+
+removeRSGroup
+superuser|global(A)
+
+
+
+removeServers
+superuser|global(A)
+
@@ -37140,7 +37247,7 @@ The server will return cellblocks compressed using this
same compressor as long
Version 3.0.0-SNAPSHOT
-Last updated 2018-01-08 14:29:39 UTC
+Last updated 2018-01-09 14:29:44 UTC
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/bulk-loads.html
--
diff --git a/bulk-loads.html b/bulk-loads.html
index cff9129..f6eae32 100644
--- a/bulk-loads.html
+++ b/bulk-loads.html
@@ -7,7 +7,7 @@
-
+
Apache HBase
Bulk Loads in Apache HBase (TM)
@@ -311,7 +311,7 @@ under the License. -->
https://www.apache.org/;>The Apache Software
Foundation.
All rights reserved.
- Last Published:
2018-01-08
+ Last Published:
2018-01-09
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c7c40c62/checkstyle-aggregate.html
--
diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html
index 236593f..c7738b4 100644
--- a/checkstyle-aggregate.html
+++ b/checkstyle-aggregate.html
@@ -7,7 +7,7 @@
-
+
Apache HBase Checkstyle Results
@@ -289,7 +289,7 @@
3471
0
0
-19048
+18828
Files
@@ -397,7 +397,7 @@
org/apache/hadoop/hbase/DistributedHBaseCluster.java
0
0
-5
+3
org/apache/hadoop/hbase/DoNotRetryIOException.java
0
@@ -422,7 +422,7 @@
org/apache/hadoop/hbase/HBaseCluster.java
0
0
-4
+2
org/apache/hadoop/hbase/HBaseClusterManager.java
0
@@ -534,1410 +534,1400 @@
0
3
-org/apache/hadoop/hbase/IntegrationTestLazyCfLoading.java
-0
-0
-2
-
org/apache/hadoop/hbase/IntegrationTestManyRegions.java
0
0
2
-
+
org/apache/hadoop/hbase/IntegrationTestMetaReplicas.java
0
0
2
-
+
org/apache/hadoop/hbase/IntegrationTestRegionReplicaPerf.java
0
0
11
-
+
org/apache/hadoop/hbase/IntegrationTestRegionReplicaReplication.java
0
0
1
-
+
org/apache/hadoop/hbase/IntegrationTestingUtility.java
0
0
5
-
+
org/apache/hadoop/hbase/JMXListener.java
0
0
5
-
+
org/apache/hadoop/hbase/KeyValue.java
0
0
117
-
+
org/apache/hadoop/hbase/KeyValueTestUtil.java
0
0
8
-
+
org/apache/hadoop/hbase/KeyValueUtil.java
0
0
29
-
+
org/apache/hadoop/hbase/LocalHBaseCluster.java
0
0
24
-
+
org/apache/hadoop/hbase/MetaMockingUtil.java
0
0
4
-
+
org/apache/hadoop/hbase/MetaMutationAnnotation.java
0
0
1
-
+
org/apache/hadoop/hbase/MetaTableAccessor.java
0
0
114
-
+
org/apache/hadoop/hbase/MiniHBaseCluster.java
0
0
-28
-
+26
+
org/apache/hadoop/hbase/MockRegionServerServices.java
0
0
1
-
+
org/apache/hadoop/hbase/MultithreadedTestUtil.java
0
0
3
-
+
org/apache/hadoop/hbase/NamespaceDescriptor.java
0
0
3
-
+
org/apache/hadoop/hbase/NotAllMetaRegionsOnlineException.java
0
0
1
-
+
org/apache/hadoop/hbase/NotServingRegionException.java
0
0
1
-
+
org/apache/hadoop/hbase/PerformanceEvaluation.java
0
0
29
-
+
org/apache/hadoop/hbase/PerformanceEvaluationCommons.java
0
0
1
-
+
org/apache/hadoop/hbase/PrivateCellUtil.java
0
0
67
-
+
org/apache/hadoop/hbase/QosTestHelper.java
0
0
3
-
+
org/apache/hadoop/hbase/RESTApiClusterManager.java
0
0
2
-
+
org/apache/hadoop/hbase/RegionLoad.java
0
0
1
-
+
org/apache/hadoop/hbase/RegionLocations.java
0
0
11
-
+
org/apache/hadoop/hbase/RegionStateListener.java
0
0
2
-
+
org/apache/hadoop/hbase/ResourceChecker.java
0
0
4
-
+
org/apache/hadoop/hbase/ScanPerformanceEvaluation.java
0
0
12
-
+
org/apache/hadoop/hbase/ScheduledChore.java
0
0
5
-
+
org/apache/hadoop/hbase/Server.java
0
0
1
-
+
org/apache/hadoop/hbase/ServerLoad.java
0
0
1
-
+
org/apache/hadoop/hbase/ServerName.java
0
0
24
-
+
org/apache/hadoop/hbase/SplitLogCounters.java
0
0
1
-
+
org/apache/hadoop/hbase/SplitLogTask.java
0
0
4
-
+
org/apache/hadoop/hbase/StripeCompactionsPerformanceEvaluation.java
0
0
-5
-
+3
+
org/apache/hadoop/hbase/TableDescriptors.java
0
0
9
-
+
org/apache/hadoop/hbase/TableInfoMissingException.java
0
0
6
-
+
org/apache/hadoop/hbase/TableName.java
0
0
17
-
+
org/apache/hadoop/hbase/TableNotDisabledException.java
0
0
1
-
+
org/apache/hadoop/hbase/TableNotEnabledException.java
0
0
1
-
+
org/apache/hadoop/hbase/TableNotFoundException.java
0
0
1
-
+
org/apache/hadoop/hbase/TagType.java
0
0
1
-
+
org/apache/hadoop/hbase/TestClassFinder.java
0
0
1
-
+
org/apache/hadoop/hbase/TestClientClusterStatus.java
0
0
1
-
+
org/apache/hadoop/hbase/TestFSTableDescriptorForceCreation.java
0
0
1
-
+
org/apache/hadoop/hbase/TestFullLogReconstruction.java
0
0
3
-
+
org/apache/hadoop/hbase/TestGlobalMemStoreSize.java
0
0
5
-
+
org/apache/hadoop/hbase/TestHBaseConfiguration.java
0
0
1
-
+
org/apache/hadoop/hbase/TestHBaseOnOtherDfsCluster.java
0
0
2
-
+
org/apache/hadoop/hbase/TestHBaseTestingUtility.java
0
0
6
-
+
org/apache/hadoop/hbase/TestHColumnDescriptor.java
0
0
1
-
+
org/apache/hadoop/hbase/TestHColumnDescriptorDefaultVersions.java
0
0
1
-
+
org/apache/hadoop/hbase/TestHDFSBlocksDistribution.java
0
0
2
-
+
org/apache/hadoop/hbase/TestHTableDescriptor.java
0
0
11
-
+
org/apache/hadoop/hbase/TestIOFencing.java
0
0
3
-
+
org/apache/hadoop/hbase/TestInfoServers.java
0
0
1
-
+
org/apache/hadoop/hbase/TestInterfaceAudienceAnnotations.java
0
0
2
-
+
org/apache/hadoop/hbase/TestJMXListener.java
0
0
1
-
+
org/apache/hadoop/hbase/TestKeyValue.java
0
0
3
-
+
org/apache/hadoop/hbase/TestLocalHBaseCluster.java
0
0
1
-
+
org/apache/hadoop/hbase/TestMetaTableAccessor.java
0
0
5
-
+
org/apache/hadoop/hbase/TestMetaTableAccessorNoCluster.java
0
0
8
-
+
org/apache/hadoop/hbase/TestMetaTableLocator.java
0
0
35
-
+
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bb398572/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsHeapMemoryManagerSource.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsHeapMemoryManagerSource.html
b/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsHeapMemoryManagerSource.html
index 1088938..a9b8e6c 100644
---
a/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsHeapMemoryManagerSource.html
+++
b/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsHeapMemoryManagerSource.html
@@ -692,7 +692,7 @@ extends
ABOVE_HEAP_LOW_WATERMARK_COUNTER_NAME
-static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String ABOVE_HEAP_LOW_WATERMARK_COUNTER_NAME
+static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String ABOVE_HEAP_LOW_WATERMARK_COUNTER_NAME
See Also:
Constant
Field Values
@@ -705,7 +705,7 @@ extends
ABOVE_HEAP_LOW_WATERMARK_COUNTER_DESC
-static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String ABOVE_HEAP_LOW_WATERMARK_COUNTER_DESC
+static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String ABOVE_HEAP_LOW_WATERMARK_COUNTER_DESC
See Also:
Constant
Field Values
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bb398572/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSource.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSource.html
b/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSource.html
index 3d7ee46..6c7ef76 100644
---
a/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSource.html
+++
b/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSource.html
@@ -4808,7 +4808,7 @@ extends Update the PutBatch time histogram if a batch contains a
Put op
Parameters:
-t -
+t - time it took
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bb398572/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSourceImpl.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSourceImpl.html
b/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSourceImpl.html
index 23a1532..ec53127 100644
---
a/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSourceImpl.html
+++
b/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsRegionServerSourceImpl.html
@@ -1615,6 +1615,8 @@ implements Specified by:
updatePutBatchin
interfaceMetricsRegionServerSource
+Parameters:
+t - time it took
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bb398572/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapper.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapper.html
b/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapper.html
index 5bf1e61..fe902d5 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapper.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/MetricsRegionWrapper.html
@@ -142,11 +142,15 @@ var activeTableTab = "activeTableTab";
long
-getMaxCompactionQueueSize()
+getMaxCompactionQueueSize()
+Note that this metric is updated periodically and hence
might miss some data points.
+
long
-getMaxFlushQueueSize()
+getMaxFlushQueueSize()
+Note that this metric is updated periodically and hence
might miss some data points.
+
long
@@ -490,8 +494,8 @@ var activeTableTab = "activeTableTab";
longgetNumCompactionsQueued()
Returns:
-the total number of compactions that are currently queued(or being
executed) at point in
- time
+the total number of compactions that are currently queued(or being
executed) at point
+ in time
@@ -504,8 +508,8 @@ var activeTableTab = "activeTableTab";
longgetNumFlushesQueued()
Returns:
-the total number of flushes currently queued(being executed) for this
region at point in
- time
+the total number of flushes currently queued(being executed) for this
region at point
+ in time
@@ -515,11 +519,11 @@ var activeTableTab = "activeTableTab";
getMaxCompactionQueueSize
-longgetMaxCompactionQueueSize()
+longgetMaxCompactionQueueSize()
+Note that this metric is updated periodically and hence
might miss some data points.
Returns:
-the max number of compactions queued for this region
- Note that this metric is updated periodically and hence might miss some data
points
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/apidocs/org/apache/hadoop/hbase/security/package-summary.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/security/package-summary.html
b/apidocs/org/apache/hadoop/hbase/security/package-summary.html
index ff9392b..8767d6f 100644
--- a/apidocs/org/apache/hadoop/hbase/security/package-summary.html
+++ b/apidocs/org/apache/hadoop/hbase/security/package-summary.html
@@ -157,6 +157,6 @@
-Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
+Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/apidocs/org/apache/hadoop/hbase/security/package-tree.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/security/package-tree.html
b/apidocs/org/apache/hadoop/hbase/security/package-tree.html
index f7ac06d..4f1e786 100644
--- a/apidocs/org/apache/hadoop/hbase/security/package-tree.html
+++ b/apidocs/org/apache/hadoop/hbase/security/package-tree.html
@@ -154,6 +154,6 @@
-Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
+Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/apidocs/org/apache/hadoop/hbase/security/package-use.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/security/package-use.html
b/apidocs/org/apache/hadoop/hbase/security/package-use.html
index 1da2a37..40d773c 100644
--- a/apidocs/org/apache/hadoop/hbase/security/package-use.html
+++ b/apidocs/org/apache/hadoop/hbase/security/package-use.html
@@ -199,6 +199,6 @@
-Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
+Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/apidocs/org/apache/hadoop/hbase/shaded/protobuf/package-summary.html
--
diff --git
a/apidocs/org/apache/hadoop/hbase/shaded/protobuf/package-summary.html
b/apidocs/org/apache/hadoop/hbase/shaded/protobuf/package-summary.html
index 4c8031a..02b82c7 100644
--- a/apidocs/org/apache/hadoop/hbase/shaded/protobuf/package-summary.html
+++ b/apidocs/org/apache/hadoop/hbase/shaded/protobuf/package-summary.html
@@ -119,6 +119,6 @@
-Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
+Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/apidocs/org/apache/hadoop/hbase/shaded/protobuf/package-tree.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/shaded/protobuf/package-tree.html
b/apidocs/org/apache/hadoop/hbase/shaded/protobuf/package-tree.html
index bc6b212..3903b67 100644
--- a/apidocs/org/apache/hadoop/hbase/shaded/protobuf/package-tree.html
+++ b/apidocs/org/apache/hadoop/hbase/shaded/protobuf/package-tree.html
@@ -123,6 +123,6 @@
-Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
+Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/apidocs/org/apache/hadoop/hbase/shaded/protobuf/package-use.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/shaded/protobuf/package-use.html
b/apidocs/org/apache/hadoop/hbase/shaded/protobuf/package-use.html
index 75fed5b..0586399 100644
--- a/apidocs/org/apache/hadoop/hbase/shaded/protobuf/package-use.html
+++ b/apidocs/org/apache/hadoop/hbase/shaded/protobuf/package-use.html
@@ -120,6 +120,6 @@
-Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
+Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/apidocs/org/apache/hadoop/hbase/snapshot/CorruptedSnapshotException.html
--
diff --git
a/apidocs/org/apache/hadoop/hbase/snapshot/CorruptedSnapshotException.html
b/apidocs/org/apache/hadoop/hbase/snapshot/CorruptedSnapshotException.html
index f05990c..83af8dd 100644
--- a/apidocs/org/apache/hadoop/hbase/snapshot/CorruptedSnapshotException.html
+++ b/apidocs/org/apache/hadoop/hbase/snapshot/CorruptedSnapshotException.html
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/83bf6175/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.html
b/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.html
index 490ee80..af3cd15 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.html
@@ -18,7 +18,7 @@
catch(err) {
}
//-->
-var methods =
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":9,"i29":10,"i30":10,"i31":10,"i32":10,"i33":9,"i34":10,"i35":9,"i36":9,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":41,"i91":41,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109"
:10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10};
+var methods =
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":9,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":9,"i35":10,"i36":9,"i37":9,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":41,"i92":41,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109"
:10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10};
var tabs = {65535:["t0","All Methods"],1:["t1","Static
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete
Methods"],32:["t6","Deprecated Methods"]};
var altColor = "altColor";
var rowColor = "rowColor";
@@ -548,7 +548,7 @@ implements
-private HStoreFile
+protected HStoreFile
createStoreFileAndReader(org.apache.hadoop.fs.Pathp)
@@ -579,10 +579,18 @@ implements determineTTLFromFamily(ColumnFamilyDescriptorfamily)
+protected http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListHStoreFile
+doCompaction(CompactionRequestImplcr,
+http://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
title="class or interface in java.util">CollectionHStoreFilefilesToCompact,
+Useruser,
+longcompactionStartTime,
+http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in
java.util">Listorg.apache.hadoop.fs.PathnewFiles)
+
+
private void
finishCompactionRequest(CompactionRequestImplcr)
-
+
protected http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in
java.util">Listorg.apache.hadoop.fs.Path
flushCache(longlogCacheFlushId,
MemStoreSnapshotsnapshot,
@@ -592,183 +600,183 @@ implements Write out current snapshot.
-
+
http://docs.oracle.com/javase/8/docs/api/java/util/OptionalDouble.html?is-external=true;
title="class or interface in java.util">OptionalDouble
getAvgStoreFileAge()
-
+
long
getBlockingFileCount()
The number of files required before flushes for this store
will be blocked.
-
+
static int
getBytesPerChecksum(org.apache.hadoop.conf.Configurationconf)
Returns the configured bytesPerChecksum value.
-
+
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/apidocs/src-html/org/apache/hadoop/hbase/io/crypto/Context.html
--
diff --git a/apidocs/src-html/org/apache/hadoop/hbase/io/crypto/Context.html
b/apidocs/src-html/org/apache/hadoop/hbase/io/crypto/Context.html
index a05dedc..bdde155 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/io/crypto/Context.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/io/crypto/Context.html
@@ -32,7 +32,7 @@
024import
org.apache.hadoop.hbase.util.MD5Hash;
025import
org.apache.yetus.audience.InterfaceAudience;
026
-027import
org.apache.hadoop.hbase.shaded.com.google.common.base.Preconditions;
+027import
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
028
029/**
030 * Crypto context. Encapsulates an
encryption algorithm and its key material.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/apidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcClientConfigHelper.html
--
diff --git
a/apidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcClientConfigHelper.html
b/apidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcClientConfigHelper.html
index a3bc373..19062d9 100644
---
a/apidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcClientConfigHelper.html
+++
b/apidocs/src-html/org/apache/hadoop/hbase/ipc/NettyRpcClientConfigHelper.html
@@ -25,10 +25,10 @@
017 */
018package org.apache.hadoop.hbase.ipc;
019
-020import
org.apache.hadoop.hbase.shaded.com.google.common.base.Preconditions;
+020import
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
021
-022import
org.apache.hadoop.hbase.shaded.io.netty.channel.Channel;
-023import
org.apache.hadoop.hbase.shaded.io.netty.channel.EventLoopGroup;
+022import
org.apache.hbase.thirdparty.io.netty.channel.Channel;
+023import
org.apache.hbase.thirdparty.io.netty.channel.EventLoopGroup;
024
025import java.util.HashMap;
026import java.util.Map;
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/CellCounter.html
--
diff --git
a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/CellCounter.html
b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/CellCounter.html
index ab79d90..67249a1 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/CellCounter.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/CellCounter.html
@@ -57,7 +57,7 @@
049import org.apache.hadoop.util.Tool;
050import
org.apache.hadoop.util.ToolRunner;
051
-052import
org.apache.hadoop.hbase.shaded.com.google.common.base.Preconditions;
+052import
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
053
054
055/**
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html
--
diff --git
a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html
b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html
index 9092878..aa4c039 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html
@@ -99,7 +99,7 @@
091import org.slf4j.Logger;
092import org.slf4j.LoggerFactory;
093
-094import
org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
+094import
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
095
096/**
097 * Writes HFiles. Passed Cells must
arrive in order.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.html
--
diff --git a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.html
b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.html
index 3b84731..464d55c 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/Import.html
@@ -650,7 +650,7 @@
642
job.setPartitionerClass(CellWritableComparablePartitioner.class);
643
job.setNumReduceTasks(regionLocator.getStartKeys().length);
644
TableMapReduceUtil.addDependencyJarsForClasses(job.getConfiguration(),
-645
org.apache.hadoop.hbase.shaded.com.google.common.base.Preconditions.class);
+645
org.apache.hbase.thirdparty.com.google.common.base.Preconditions.class);
646 }
647} else if (hfileOutPath != null) {
648 LOG.info("writing to hfiles for
bulk load.");
@@ -665,7 +665,7 @@
657
job.setMapOutputValueClass(MapReduceExtendedCell.class);
658
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d449e87f/devapidocs/org/apache/hadoop/hbase/ByteBufferKeyOnlyKeyValue.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/ByteBufferKeyOnlyKeyValue.html
b/devapidocs/org/apache/hadoop/hbase/ByteBufferKeyOnlyKeyValue.html
index de2a7ac..6e471c6 100644
--- a/devapidocs/org/apache/hadoop/hbase/ByteBufferKeyOnlyKeyValue.html
+++ b/devapidocs/org/apache/hadoop/hbase/ByteBufferKeyOnlyKeyValue.html
@@ -18,7 +18,7 @@
catch(err) {
}
//-->
-var methods =
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10};
+var methods =
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10};
var tabs = {65535:["t0","All Methods"],2:["t2","Instance
Methods"],8:["t4","Concrete Methods"]};
var altColor = "altColor";
var rowColor = "rowColor";
@@ -49,7 +49,7 @@ var activeTableTab = "activeTableTab";
-PrevClass
+PrevClass
NextClass
@@ -100,7 +100,7 @@ var activeTableTab = "activeTableTab";
http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">java.lang.Object
-org.apache.hadoop.hbase.ByteBufferCell
+org.apache.hadoop.hbase.ByteBufferExtendedCell
org.apache.hadoop.hbase.ByteBufferKeyOnlyKeyValue
@@ -114,13 +114,13 @@ var activeTableTab = "activeTableTab";
All Implemented Interfaces:
-Cell
+http://docs.oracle.com/javase/8/docs/api/java/lang/Cloneable.html?is-external=true;
title="class or interface in java.lang">Cloneable, Cell, ExtendedCell, HeapSize, RawCell
@InterfaceAudience.Private
-public class ByteBufferKeyOnlyKeyValue
-extends ByteBufferCell
+public class ByteBufferKeyOnlyKeyValue
+extends ByteBufferExtendedCell
This is a key only Cell implementation which is identical
to KeyValue.KeyOnlyKeyValue
with respect to key serialization but have its data in the form of Byte buffer
(onheap and offheap).
@@ -141,7 +141,7 @@ extends
Nested classes/interfaces inherited from
interfaceorg.apache.hadoop.hbase.Cell
-Cell.DataType
+Cell.Type
@@ -162,18 +162,36 @@ extends buf
+static int
+FIXED_OVERHEAD
+
+
private int
length
-
+
private int
offset
-
+
private short
rowLen
+
+
+
+
+Fields inherited from interfaceorg.apache.hadoop.hbase.ExtendedCell
+CELL_NOT_BASED_ON_CHUNK
+
+
+
+
+
+Fields inherited from interfaceorg.apache.hadoop.hbase.RawCell
+MAX_TAGS_LENGTH
+
@@ -300,70 +318,86 @@ extends
+http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true;
title="class or interface in java.util">OptionalTag
+getTag(bytetype)
+Returns the specific tag of the given type
+
+
+
+http://docs.oracle.com/javase/8/docs/api/java/util/Iterator.html?is-external=true;
title="class or interface in java.util">IteratorTag
+getTags()
+Creates a list of tags in the current cell
+
+
+
byte[]
getTagsArray()
Contiguous raw bytes representing tags that may start at
any index in the containing array.
-
+
http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
title="class or interface in java.nio">ByteBuffer
getTagsByteBuffer()
-
+
int
getTagsLength()
HBase internally uses 2 bytes to store tags length in
Cell.
-
+
int
getTagsOffset()
-
+
int
getTagsPosition()
-
+
long
getTimestamp()
-
+
private int
getTimestampOffset()
-
-Cell.DataType
+
+Cell.Type
getType()
-Returns the type of cell in a human readable format using
Cell.DataType
+Returns the type of cell in a human readable format using
Cell.Type
-
+
byte
getTypeByte()
-
+
byte[]
getValueArray()
Contiguous raw bytes that may start at any index in the
containing array.
-
+
http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
title="class or interface in java.nio">ByteBuffer
getValueByteBuffer()
-
+
int
getValueLength()
-
+
int
getValueOffset()
-
+
int
getValuePosition()
-
+
+long
+heapSize()
+
+
void
setKey(http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
title="class or interface in java.nio">ByteBufferkey,
intoffset,
@@ -372,7 +406,25 @@ extends
-
+
+void
+setSequenceId(longseqId)
+Sets with the given seqId.
+
+
+
+void
+setTimestamp(byte[]ts)
+Sets with the given timestamp.
+
+
+
+void
+setTimestamp(longts)
+Sets with the given timestamp.
+
+
+
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d2b28a1a/devapidocs/org/apache/hadoop/hbase/KeyValue.KeyOnlyKeyValue.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/KeyValue.KeyOnlyKeyValue.html
b/devapidocs/org/apache/hadoop/hbase/KeyValue.KeyOnlyKeyValue.html
index 321a54e..6636989 100644
--- a/devapidocs/org/apache/hadoop/hbase/KeyValue.KeyOnlyKeyValue.html
+++ b/devapidocs/org/apache/hadoop/hbase/KeyValue.KeyOnlyKeyValue.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
-public static class KeyValue.KeyOnlyKeyValue
+public static class KeyValue.KeyOnlyKeyValue
extends KeyValue
A simple form of KeyValue that creates a keyvalue with only
the key part of the byte[]
Mainly used in places where we need to compare two cells. Avoids copying of
bytes
@@ -394,7 +394,7 @@ extends KeyValue
-checkParameters,
clone,
create,
create,
createKeyOnly,
deepClone,
getBuffer,
getDelimiter,
getDelimiterInReverse,
getFamilyLength, getKeyDataStructureSize,
getKeyString,
getKeyValueDataStructureSize,
getKeyValueDataStructureSize,
getKeyValueDataStructureSize,
getLength,
getOffset,
getSequenceId,
getSerializedSize, getTag,
getTags,
getTimestamp,
humanReadableTimestamp,
isLatestTimestamp,
keyToString,
keyToString,
oswrite,
setSequenceId,
setTimestamp,
setTimestamp,
shallowCopy,
toStringMap,
updateLatestStamp,
write,
write,
writeByteArray
+checkParameters,
clone,
create,
create,
createKeyOnly,
deepClone,
getBuffer,
getDelimiter,
getDelimiterInReverse,
getFamilyLength, getKeyDataStructureSize,
getKeyString,
getKeyValueDataStructureSize,
getKeyValueDataStructureSize,
getKeyValueDataStructureSize,
getLength,
getOffset,
getSequenceId,
getSerializedSize, getTimestamp,
humanReadableTimestamp,
isLatestTimestamp,
keyToString,
keyToString,
oswrite,
setSequenceId,
setTimestamp,
set
Timestamp, shallowCopy,
toStringMap,
updateLatestStamp,
write,
write,
writeByteArray
@@ -415,7 +415,7 @@ extends RawCell
-checkForTagsLength,
cloneTags
+checkForTagsLength,
cloneTags,
getTag,
getTags
@@ -437,7 +437,7 @@ extends
rowLen
-privateshort rowLen
+privateshort rowLen
@@ -454,7 +454,7 @@ extends
KeyOnlyKeyValue
-publicKeyOnlyKeyValue()
+publicKeyOnlyKeyValue()
@@ -463,7 +463,7 @@ extends
KeyOnlyKeyValue
-publicKeyOnlyKeyValue(byte[]b)
+publicKeyOnlyKeyValue(byte[]b)
@@ -472,7 +472,7 @@ extends
KeyOnlyKeyValue
-publicKeyOnlyKeyValue(byte[]b,
+publicKeyOnlyKeyValue(byte[]b,
intoffset,
intlength)
@@ -491,7 +491,7 @@ extends
set
-publicvoidset(KeyValue.KeyOnlyKeyValuekeyOnlyKeyValue)
+publicvoidset(KeyValue.KeyOnlyKeyValuekeyOnlyKeyValue)
@@ -500,7 +500,7 @@ extends
clear
-publicvoidclear()
+publicvoidclear()
@@ -509,7 +509,7 @@ extends
getKeyOffset
-publicintgetKeyOffset()
+publicintgetKeyOffset()
Overrides:
getKeyOffsetin
classKeyValue
@@ -524,7 +524,7 @@ extends
setKey
-publicvoidsetKey(byte[]key,
+publicvoidsetKey(byte[]key,
intoffset,
intlength)
A setter that helps to avoid object creation every time and
whenever
@@ -543,7 +543,7 @@ extends
getKey
-publicbyte[]getKey()
+publicbyte[]getKey()
Description copied from
class:KeyValue
Do not use unless you have to. Used internally for
compacting and testing. Use
KeyValue.getRowArray(),
KeyValue.getFamilyArray(),
KeyValue.getQualifierArray(),
and
@@ -562,7 +562,7 @@ extends
getRowArray
-publicbyte[]getRowArray()
+publicbyte[]getRowArray()
Description copied from
interface:Cell
Contiguous raw bytes that may start at any index in the
containing array. Max length is
Short.MAX_VALUE which is 32,767 bytes.
@@ -582,7 +582,7 @@ extends
getRowOffset
-publicintgetRowOffset()
+publicintgetRowOffset()
Specified by:
getRowOffsetin
interfaceCell
@@ -599,7 +599,7 @@ extends
getFamilyArray
-publicbyte[]getFamilyArray()
+publicbyte[]getFamilyArray()
Description copied from
interface:Cell
Contiguous bytes composed of legal HDFS filename characters
which may start at any index in the
containing array. Max length is Byte.MAX_VALUE, which is 127 bytes.
@@ -619,7 +619,7 @@ extends
getFamilyLength
-publicbytegetFamilyLength()
+publicbytegetFamilyLength()
Specified by:
getFamilyLengthin
interfaceCell
@@ -636,7 +636,7 @@ extends
getFamilyOffset
-publicintgetFamilyOffset()
+publicintgetFamilyOffset()
Specified by:
getFamilyOffsetin
interfaceCell
@@ -653,7 +653,7 @@ extends
getQualifierArray
-publicbyte[]getQualifierArray()
+publicbyte[]getQualifierArray()
Description copied from
interface:Cell
Contiguous raw bytes that may start at any index in the
containing array.
@@ -672,7 +672,7 @@ extends
getQualifierLength
-publicintgetQualifierLength()
+publicintgetQualifierLength()
Specified by:
getQualifierLengthin
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b618ac40/apidocs/src-html/org/apache/hadoop/hbase/util/Bytes.ByteArrayComparator.html
--
diff --git
a/apidocs/src-html/org/apache/hadoop/hbase/util/Bytes.ByteArrayComparator.html
b/apidocs/src-html/org/apache/hadoop/hbase/util/Bytes.ByteArrayComparator.html
index 86fc15e..d02bcdf 100644
---
a/apidocs/src-html/org/apache/hadoop/hbase/util/Bytes.ByteArrayComparator.html
+++
b/apidocs/src-html/org/apache/hadoop/hbase/util/Bytes.ByteArrayComparator.html
@@ -38,2586 +38,2590 @@
030import java.nio.ByteBuffer;
031import
java.nio.charset.StandardCharsets;
032import java.security.SecureRandom;
-033import java.util.Arrays;
-034import java.util.Collection;
-035import java.util.Comparator;
-036import java.util.Iterator;
-037import java.util.List;
-038
-039import org.apache.hadoop.hbase.Cell;
-040import
org.apache.hadoop.hbase.CellComparator;
-041import
org.apache.hadoop.hbase.KeyValue;
-042import
org.apache.hadoop.io.RawComparator;
-043import
org.apache.hadoop.io.WritableComparator;
-044import
org.apache.hadoop.io.WritableUtils;
-045import
org.apache.yetus.audience.InterfaceAudience;
-046import org.slf4j.Logger;
-047import org.slf4j.LoggerFactory;
-048
-049import
org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
-050import
org.apache.hadoop.hbase.shaded.com.google.common.collect.Lists;
-051
-052import com.google.protobuf.ByteString;
-053
-054import sun.misc.Unsafe;
-055
-056/**
-057 * Utility class that handles byte
arrays, conversions to/from other types,
-058 * comparisons, hash code generation,
manufacturing keys for HashMaps or
-059 * HashSets, and can be used as key in
maps or trees.
-060 */
-061@SuppressWarnings("restriction")
-062@InterfaceAudience.Public
-063@edu.umd.cs.findbugs.annotations.SuppressWarnings(
-064
value="EQ_CHECK_FOR_OPERAND_NOT_COMPATIBLE_WITH_THIS",
-065justification="It has been like this
forever")
-066public class Bytes implements
ComparableBytes {
-067
-068 // Using the charset canonical name for
String/byte[] conversions is much
-069 // more efficient due to use of cached
encoders/decoders.
-070 private static final String UTF8_CSN =
StandardCharsets.UTF_8.name();
-071
-072 //HConstants.EMPTY_BYTE_ARRAY should be
updated if this changed
-073 private static final byte []
EMPTY_BYTE_ARRAY = new byte [0];
-074
-075 private static final Logger LOG =
LoggerFactory.getLogger(Bytes.class);
-076
-077 /**
-078 * Size of boolean in bytes
-079 */
-080 public static final int SIZEOF_BOOLEAN
= Byte.SIZE / Byte.SIZE;
-081
-082 /**
-083 * Size of byte in bytes
-084 */
-085 public static final int SIZEOF_BYTE =
SIZEOF_BOOLEAN;
-086
-087 /**
-088 * Size of char in bytes
-089 */
-090 public static final int SIZEOF_CHAR =
Character.SIZE / Byte.SIZE;
-091
-092 /**
-093 * Size of double in bytes
-094 */
-095 public static final int SIZEOF_DOUBLE =
Double.SIZE / Byte.SIZE;
-096
-097 /**
-098 * Size of float in bytes
-099 */
-100 public static final int SIZEOF_FLOAT =
Float.SIZE / Byte.SIZE;
-101
-102 /**
-103 * Size of int in bytes
-104 */
-105 public static final int SIZEOF_INT =
Integer.SIZE / Byte.SIZE;
-106
-107 /**
-108 * Size of long in bytes
-109 */
-110 public static final int SIZEOF_LONG =
Long.SIZE / Byte.SIZE;
-111
-112 /**
-113 * Size of short in bytes
-114 */
-115 public static final int SIZEOF_SHORT =
Short.SIZE / Byte.SIZE;
-116
-117 /**
-118 * Mask to apply to a long to reveal
the lower int only. Use like this:
-119 * int i = (int)(0xL ^
some_long_value);
-120 */
-121 public static final long
MASK_FOR_LOWER_INT_IN_LONG = 0xL;
-122
-123 /**
-124 * Estimate of size cost to pay beyond
payload in jvm for instance of byte [].
-125 * Estimate based on study of jhat and
jprofiler numbers.
-126 */
-127 // JHat says BU is 56 bytes.
-128 // SizeOf which uses
java.lang.instrument says 24 bytes. (3 longs?)
-129 public static final int
ESTIMATED_HEAP_TAX = 16;
-130
-131 private static final boolean
UNSAFE_UNALIGNED = UnsafeAvailChecker.unaligned();
-132
-133 /**
-134 * Returns length of the byte array,
returning 0 if the array is null.
-135 * Useful for calculating sizes.
-136 * @param b byte array, which can be
null
-137 * @return 0 if b is null, otherwise
returns length
-138 */
-139 final public static int len(byte[] b)
{
-140return b == null ? 0 : b.length;
-141 }
-142
-143 private byte[] bytes;
-144 private int offset;
-145 private int length;
-146
-147 /**
-148 * Create a zero-size sequence.
-149 */
-150 public Bytes() {
-151super();
-152 }
-153
-154 /**
-155 * Create a Bytes using the byte array
as the initial value.
-156 * @param bytes This array becomes the
backing storage for the object.
-157 */
-158 public Bytes(byte[] bytes) {
-159this(bytes,
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/index-all.html
--
diff --git a/devapidocs/index-all.html b/devapidocs/index-all.html
index d6491d4..a10fb33 100644
--- a/devapidocs/index-all.html
+++ b/devapidocs/index-all.html
@@ -3171,19 +3171,19 @@
Deprecated.
-appendReplicationPeerTableCFs(String,
MapTableName, ? extends CollectionString) - Method
in interface org.apache.hadoop.hbase.client.Admin
+appendReplicationPeerTableCFs(String,
MapTableName, ListString) - Method in interface
org.apache.hadoop.hbase.client.Admin
Append the replicable table column family config from the
specified peer.
-appendReplicationPeerTableCFs(String,
MapTableName, ? extends CollectionString) - Method
in interface org.apache.hadoop.hbase.client.AsyncAdmin
+appendReplicationPeerTableCFs(String,
MapTableName, ListString) - Method in interface
org.apache.hadoop.hbase.client.AsyncAdmin
Append the replicable table-cf config of the specified
peer
-appendReplicationPeerTableCFs(String,
MapTableName, ? extends CollectionString) - Method
in class org.apache.hadoop.hbase.client.AsyncHBaseAdmin
+appendReplicationPeerTableCFs(String,
MapTableName, ListString) - Method in class
org.apache.hadoop.hbase.client.AsyncHBaseAdmin
-appendReplicationPeerTableCFs(String,
MapTableName, ? extends CollectionString) - Method
in class org.apache.hadoop.hbase.client.HBaseAdmin
+appendReplicationPeerTableCFs(String,
MapTableName, ListString) - Method in class
org.apache.hadoop.hbase.client.HBaseAdmin
-appendReplicationPeerTableCFs(String,
MapTableName, ? extends CollectionString) - Method
in class org.apache.hadoop.hbase.client.RawAsyncHBaseAdmin
+appendReplicationPeerTableCFs(String,
MapTableName, ListString) - Method in class
org.apache.hadoop.hbase.client.RawAsyncHBaseAdmin
appendRSZnodes(ZKWatcher,
String, StringBuilder) - Static method in class
org.apache.hadoop.hbase.zookeeper.ZKUtil
@@ -3191,7 +3191,7 @@
appendStrings(StringBuilder,
String...) - Static method in class
org.apache.hadoop.hbase.procedure2.util.StringUtils
-appendTableCFsToReplicationPeerConfig(MapTableName,
? extends CollectionString, ReplicationPeerConfig) -
Static method in class org.apache.hadoop.hbase.client.replication.ReplicationPeerConfigUtil
+appendTableCFsToReplicationPeerConfig(MapTableName,
ListString, ReplicationPeerConfig) - Static method in
class org.apache.hadoop.hbase.client.replication.ReplicationPeerConfigUtil
appendThreadInfo(StringBuilder,
ThreadInfo, String) - Static method in class
org.apache.hadoop.hbase.monitoring.ThreadMonitoring
@@ -5244,6 +5244,8 @@
bandwidth
- Variable in class org.apache.hadoop.hbase.replication.ReplicationPeerConfig
+bandwidth
- Variable in class org.apache.hadoop.hbase.replication.ReplicationPeerConfig.ReplicationPeerConfigBuilderImpl
+
BANDWIDTH
- Static variable in class org.apache.hadoop.hbase.snapshot.ExportSnapshot.Options
bandwidthMB
- Variable in class org.apache.hadoop.hbase.snapshot.ExportSnapshot
@@ -6957,6 +6959,10 @@
Create a store file writer.
+build()
- Method in class org.apache.hadoop.hbase.replication.ReplicationPeerConfig.ReplicationPeerConfigBuilderImpl
+
+build()
- Method in interface org.apache.hadoop.hbase.replication.ReplicationPeerConfigBuilder
+
build()
- Method in class org.apache.hadoop.hbase.rest.model.ScannerModel.FilterModel
build()
- Method in class org.apache.hadoop.hbase.rest.model.ScannerModel.FilterModel.ByteArrayComparableModel
@@ -13053,6 +13059,8 @@
clusterKey
- Variable in class org.apache.hadoop.hbase.replication.ReplicationPeerConfig
+clusterKey
- Variable in class org.apache.hadoop.hbase.replication.ReplicationPeerConfig.ReplicationPeerConfigBuilderImpl
+
ClusterLoadState - Class in org.apache.hadoop.hbase.master.balancer
Class used to hold the current state of the cluster and how
balanced it is.
@@ -16060,6 +16068,8 @@
configuration
- Variable in class org.apache.hadoop.hbase.replication.ReplicationPeerConfig
+configuration
- Variable in class org.apache.hadoop.hbase.replication.ReplicationPeerConfig.ReplicationPeerConfigBuilderImpl
+
configuration
- Variable in class org.apache.hadoop.hbase.zookeeper.MiniZooKeeperCluster
ConfigurationManager - Class in org.apache.hadoop.hbase.conf
@@ -17671,10 +17681,16 @@
CopyTable()
- Constructor for class org.apache.hadoop.hbase.mapreduce.CopyTable
+copyTableCFs(MapTableName,
? extends CollectionString) - Method in class
org.apache.hadoop.hbase.client.replication.ReplicationAdmin
+
+Deprecated.
+
copyTableCFs()
- Method in class org.apache.hadoop.hbase.replication.master.ReplicationPeerConfigUpgrader
copyTableCFs(String)
- Method in class org.apache.hadoop.hbase.replication.master.ReplicationPeerConfigUpgrader
+copyTableCFsMap(MapTableName,
ListString) - Static method in class
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4b2cc17/apidocs/org/apache/hadoop/hbase/ServerMetrics.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/ServerMetrics.html
b/apidocs/org/apache/hadoop/hbase/ServerMetrics.html
new file mode 100644
index 000..170d5af
--- /dev/null
+++ b/apidocs/org/apache/hadoop/hbase/ServerMetrics.html
@@ -0,0 +1,421 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+
+
+
+ServerMetrics (Apache HBase 3.0.0-SNAPSHOT API)
+
+
+
+
+
+var methods =
{"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6};
+var tabs = {65535:["t0","All Methods"],2:["t2","Instance
Methods"],4:["t3","Abstract Methods"]};
+var altColor = "altColor";
+var rowColor = "rowColor";
+var tableTab = "tableTab";
+var activeTableTab = "activeTableTab";
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+PrevClass
+NextClass
+
+
+Frames
+NoFrames
+
+
+AllClasses
+
+
+
+
+
+
+
+Summary:
+Nested|
+Field|
+Constr|
+Method
+
+
+Detail:
+Field|
+Constr|
+Method
+
+
+
+
+
+
+
+
+org.apache.hadoop.hbase
+Interface ServerMetrics
+
+
+
+
+
+
+All Known Implementing Classes:
+ServerLoad
+
+
+
+@InterfaceAudience.Public
+public interface ServerMetrics
+This class is used for exporting current state of load on a
RegionServer.
+
+
+
+
+
+
+
+
+
+
+
+Method Summary
+
+All MethodsInstance MethodsAbstract Methods
+
+Modifier and Type
+Method and Description
+
+
+http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
+getCoprocessorNames()
+Return the RegionServer-level and Region-level
coprocessors
+
+
+
+int
+getInfoServerPort()
+
+
+long
+getLastReportTimestamp()
+
+
+Size
+getMaxHeapSize()
+
+
+http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
title="class or interface in java.util">Mapbyte[],RegionMetrics
+getRegionMetrics()
+
+
+ReplicationLoadSink
+getReplicationLoadSink()
+Call directly from client such as hbase shell
+
+
+
+http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListReplicationLoadSource
+getReplicationLoadSourceList()
+Call directly from client such as hbase shell
+
+
+
+long
+getReportTimestamp()
+
+
+long
+getRequestCount()
+
+
+long
+getRequestCountPerSecond()
+
+
+ServerName
+getServerName()
+
+
+Size
+getUsedHeapSize()
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+Method Detail
+
+
+
+
+
+getServerName
+ServerNamegetServerName()
+
+
+
+
+
+
+
+getRequestCountPerSecond
+longgetRequestCountPerSecond()
+
+Returns:
+the number of requests per second.
+
+
+
+
+
+
+
+
+getRequestCount
+longgetRequestCount()
+
+Returns:
+total Number of requests from the start of the region server.
+
+
+
+
+
+
+
+
+getUsedHeapSize
+SizegetUsedHeapSize()
+
+Returns:
+the amount of used heap
+
+
+
+
+
+
+
+
+getMaxHeapSize
+SizegetMaxHeapSize()
+
+Returns:
+the maximum allowable size of the heap
+
+
+
+
+
+
+
+
+getInfoServerPort
+intgetInfoServerPort()
+
+
+
+
+
+
+
+getReplicationLoadSourceList
+http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListReplicationLoadSourcegetReplicationLoadSourceList()
+Call directly from client such as hbase shell
+
+Returns:
+the list of ReplicationLoadSource
+
+
+
+
+
+
+
+
+getReplicationLoadSink
+@Nullable
+ReplicationLoadSinkgetReplicationLoadSink()
+Call directly from client such as hbase shell
+
+Returns:
+ReplicationLoadSink
+
+
+
+
+
+
+
+
+getRegionMetrics
+http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
title="class or interface in java.util">Mapbyte[],RegionMetricsgetRegionMetrics()
+
+Returns:
+region load metrics
+
+
+
+
+
+
+
+
+getCoprocessorNames
+http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">Listhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">StringgetCoprocessorNames()
+Return the RegionServer-level and Region-level
coprocessors
+
+Returns:
+string list of loaded RegionServer-level and Region-level coprocessors
+
+
+
+
+
+
+
+
+getReportTimestamp
+longgetReportTimestamp()
+
+Returns:
+the timestamp (server side) of
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/505bbb2e/apidocs/src-html/org/apache/hadoop/hbase/client/HTableMultiplexer.HTableMultiplexerStatus.html
--
diff --git
a/apidocs/src-html/org/apache/hadoop/hbase/client/HTableMultiplexer.HTableMultiplexerStatus.html
b/apidocs/src-html/org/apache/hadoop/hbase/client/HTableMultiplexer.HTableMultiplexerStatus.html
index 7ab77ff..3a23125 100644
---
a/apidocs/src-html/org/apache/hadoop/hbase/client/HTableMultiplexer.HTableMultiplexerStatus.html
+++
b/apidocs/src-html/org/apache/hadoop/hbase/client/HTableMultiplexer.HTableMultiplexerStatus.html
@@ -43,634 +43,633 @@
035import
java.util.concurrent.atomic.AtomicInteger;
036import
java.util.concurrent.atomic.AtomicLong;
037
-038import org.apache.commons.logging.Log;
-039import
org.apache.commons.logging.LogFactory;
-040import
org.apache.hadoop.conf.Configuration;
-041import
org.apache.hadoop.hbase.HBaseConfiguration;
-042import
org.apache.hadoop.hbase.HConstants;
-043import
org.apache.hadoop.hbase.HRegionLocation;
-044import
org.apache.hadoop.hbase.ServerName;
-045import
org.apache.hadoop.hbase.TableName;
-046import
org.apache.hadoop.hbase.ipc.RpcControllerFactory;
-047import
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-048import
org.apache.yetus.audience.InterfaceAudience;
-049
-050import
org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
-051import
org.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.ThreadFactoryBuilder;
-052
-053/**
-054 * HTableMultiplexer provides a
thread-safe non blocking PUT API across all the tables.
-055 * Each put will be sharded into
different buffer queues based on its destination region server.
-056 * So each region server buffer queue
will only have the puts which share the same destination.
-057 * And each queue will have a flush
worker thread to flush the puts request to the region server.
-058 * If any queue is full, the
HTableMultiplexer starts to drop the Put requests for that
-059 * particular queue.
-060 *
-061 * Also all the puts will be retried as a
configuration number before dropping.
-062 * And the HTableMultiplexer can report
the number of buffered requests and the number of the
-063 * failed (dropped) requests in total or
on per region server basis.
-064 *
-065 * This class is thread safe.
-066 */
-067@InterfaceAudience.Public
-068public class HTableMultiplexer {
-069 private static final Log LOG =
LogFactory.getLog(HTableMultiplexer.class.getName());
-070
-071 public static final String
TABLE_MULTIPLEXER_FLUSH_PERIOD_MS =
-072
"hbase.tablemultiplexer.flush.period.ms";
-073 public static final String
TABLE_MULTIPLEXER_INIT_THREADS = "hbase.tablemultiplexer.init.threads";
-074 public static final String
TABLE_MULTIPLEXER_MAX_RETRIES_IN_QUEUE =
-075
"hbase.client.max.retries.in.queue";
-076
-077 /** The map between each region server
to its flush worker */
-078 private final MapHRegionLocation,
FlushWorker serverToFlushWorkerMap =
-079 new ConcurrentHashMap();
-080
-081 private final Configuration
workerConf;
-082 private final ClusterConnection conn;
-083 private final ExecutorService pool;
-084 private final int maxAttempts;
-085 private final int
perRegionServerBufferQueueSize;
-086 private final int maxKeyValueSize;
-087 private final ScheduledExecutorService
executor;
-088 private final long flushPeriod;
-089
-090 /**
-091 * @param conf The HBaseConfiguration
-092 * @param
perRegionServerBufferQueueSize determines the max number of the buffered Put
ops for
-093 * each region server before
dropping the request.
-094 */
-095 public HTableMultiplexer(Configuration
conf, int perRegionServerBufferQueueSize)
-096 throws IOException {
-097
this(ConnectionFactory.createConnection(conf), conf,
perRegionServerBufferQueueSize);
-098 }
-099
-100 /**
-101 * @param conn The HBase connection.
-102 * @param conf The HBase
configuration
-103 * @param
perRegionServerBufferQueueSize determines the max number of the buffered Put
ops for
-104 * each region server before
dropping the request.
-105 */
-106 public HTableMultiplexer(Connection
conn, Configuration conf,
-107 int perRegionServerBufferQueueSize)
{
-108this.conn = (ClusterConnection)
conn;
-109this.pool =
HTable.getDefaultExecutor(conf);
-110// how many times we could try in
total, one more than retry number
-111this.maxAttempts =
conf.getInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER,
-112
HConstants.DEFAULT_HBASE_CLIENT_RETRIES_NUMBER) + 1;
-113this.perRegionServerBufferQueueSize =
perRegionServerBufferQueueSize;
-114this.maxKeyValueSize =
HTable.getMaxKeyValueSize(conf);
-115this.flushPeriod =
conf.getLong(TABLE_MULTIPLEXER_FLUSH_PERIOD_MS, 100);
-116int initThreads =
conf.getInt(TABLE_MULTIPLEXER_INIT_THREADS, 10);
-117
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4c0cfa5/apidocs/org/apache/hadoop/hbase/types/OrderedFloat32.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/types/OrderedFloat32.html
b/apidocs/org/apache/hadoop/hbase/types/OrderedFloat32.html
index a56d88b..a869c33 100644
--- a/apidocs/org/apache/hadoop/hbase/types/OrderedFloat32.html
+++ b/apidocs/org/apache/hadoop/hbase/types/OrderedFloat32.html
@@ -119,7 +119,7 @@ var activeTableTab = "activeTableTab";
@InterfaceAudience.Public
-public class OrderedFloat32
+public class OrderedFloat32
extends OrderedBytesBasehttp://docs.oracle.com/javase/8/docs/api/java/lang/Float.html?is-external=true;
title="class or interface in java.lang">Float
A float of 32-bits using a fixed-length
encoding. Based on
OrderedBytes.encodeFloat32(PositionedByteRange,
float, Order).
@@ -269,7 +269,7 @@ extends
ASCENDING
-public static finalOrderedFloat32 ASCENDING
+public static finalOrderedFloat32 ASCENDING
@@ -278,7 +278,7 @@ extends
DESCENDING
-public static finalOrderedFloat32 DESCENDING
+public static finalOrderedFloat32 DESCENDING
@@ -295,7 +295,7 @@ extends
OrderedFloat32
-protectedOrderedFloat32(Orderorder)
+protectedOrderedFloat32(Orderorder)
@@ -312,7 +312,7 @@ extends
isNullable
-publicbooleanisNullable()
+publicbooleanisNullable()
Description copied from
interface:DataType
Indicates whether this instance supports encoding null
values. This
depends on the implementation details of the encoding format. All
@@ -334,7 +334,7 @@ extends
encodedLength
-publicintencodedLength(http://docs.oracle.com/javase/8/docs/api/java/lang/Float.html?is-external=true;
title="class or interface in java.lang">Floatval)
+publicintencodedLength(http://docs.oracle.com/javase/8/docs/api/java/lang/Float.html?is-external=true;
title="class or interface in java.lang">Floatval)
Description copied from
interface:DataType
Inform consumers how long the encoded byte[]
will be.
@@ -351,7 +351,7 @@ extends
encodedClass
-publichttp://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true;
title="class or interface in java.lang">Classhttp://docs.oracle.com/javase/8/docs/api/java/lang/Float.html?is-external=true;
title="class or interface in java.lang">FloatencodedClass()
+publichttp://docs.oracle.com/javase/8/docs/api/java/lang/Class.html?is-external=true;
title="class or interface in java.lang">Classhttp://docs.oracle.com/javase/8/docs/api/java/lang/Float.html?is-external=true;
title="class or interface in java.lang">FloatencodedClass()
Description copied from
interface:DataType
Inform consumers over what type this DataType
operates. Useful
when working with bare DataType instances.
@@ -363,7 +363,7 @@ extends
decode
-publichttp://docs.oracle.com/javase/8/docs/api/java/lang/Float.html?is-external=true;
title="class or interface in java.lang">Floatdecode(PositionedByteRangesrc)
+publichttp://docs.oracle.com/javase/8/docs/api/java/lang/Float.html?is-external=true;
title="class or interface in java.lang">Floatdecode(PositionedByteRangesrc)
Description copied from
interface:DataType
Read an instance of T from the buffer
src.
@@ -378,7 +378,7 @@ extends
encode
-publicintencode(PositionedByteRangedst,
+publicintencode(PositionedByteRangedst,
http://docs.oracle.com/javase/8/docs/api/java/lang/Float.html?is-external=true;
title="class or interface in java.lang">Floatval)
Description copied from
interface:DataType
Write instance val into buffer
dst.
@@ -397,7 +397,7 @@ extends
decodeFloat
-publicfloatdecodeFloat(PositionedByteRangedst)
+publicfloatdecodeFloat(PositionedByteRangedst)
Read a float value from the buffer
dst.
@@ -407,7 +407,7 @@ extends
encodeFloat
-publicintencodeFloat(PositionedByteRangedst,
+publicintencodeFloat(PositionedByteRangedst,
floatval)
Write instance val into buffer
buff.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4c0cfa5/apidocs/org/apache/hadoop/hbase/types/OrderedFloat64.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/types/OrderedFloat64.html
b/apidocs/org/apache/hadoop/hbase/types/OrderedFloat64.html
index 9d77ff6..4fbac85 100644
--- a/apidocs/org/apache/hadoop/hbase/types/OrderedFloat64.html
+++ b/apidocs/org/apache/hadoop/hbase/types/OrderedFloat64.html
@@ -119,7 +119,7 @@ var activeTableTab = "activeTableTab";
@InterfaceAudience.Public
-public class OrderedFloat64
+public class OrderedFloat64
extends OrderedBytesBasehttp://docs.oracle.com/javase/8/docs/api/java/lang/Double.html?is-external=true;
title="class or interface in java.lang">Double
A double of 64-bits using a fixed-length
encoding. Built on
OrderedBytes.encodeFloat64(PositionedByteRange,
double, Order).
@@ -269,7 +269,7 @@ extends
ASCENDING
-public static
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/071f974b/devapidocs/org/apache/hadoop/hbase/client/example/MultiThreadedClientExample.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/client/example/MultiThreadedClientExample.html
b/devapidocs/org/apache/hadoop/hbase/client/example/MultiThreadedClientExample.html
index 2194a60..e4e6ccf 100644
---
a/devapidocs/org/apache/hadoop/hbase/client/example/MultiThreadedClientExample.html
+++
b/devapidocs/org/apache/hadoop/hbase/client/example/MultiThreadedClientExample.html
@@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab";
-public class MultiThreadedClientExample
+public class MultiThreadedClientExample
extends org.apache.hadoop.conf.Configured
implements org.apache.hadoop.util.Tool
Example on how to use HBase's Connection
and Table in a
@@ -321,7 +321,7 @@ implements org.apache.hadoop.util.Tool
LOG
-private static finalorg.apache.commons.logging.Log LOG
+private static finalorg.apache.commons.logging.Log LOG
@@ -330,7 +330,7 @@ implements org.apache.hadoop.util.Tool
DEFAULT_NUM_OPERATIONS
-private static finalint DEFAULT_NUM_OPERATIONS
+private static finalint DEFAULT_NUM_OPERATIONS
See Also:
Constant
Field Values
@@ -343,7 +343,7 @@ implements org.apache.hadoop.util.Tool
FAMILY
-private static finalbyte[] FAMILY
+private static finalbyte[] FAMILY
The name of the column family.
d for default.
@@ -355,7 +355,7 @@ implements org.apache.hadoop.util.Tool
QUAL
-private static finalbyte[] QUAL
+private static finalbyte[] QUAL
For the example we're just using one qualifier.
@@ -365,7 +365,7 @@ implements org.apache.hadoop.util.Tool
internalPool
-private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true;
title="class or interface in java.util.concurrent">ExecutorService internalPool
+private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true;
title="class or interface in java.util.concurrent">ExecutorService internalPool
@@ -374,7 +374,7 @@ implements org.apache.hadoop.util.Tool
threads
-private finalint threads
+private finalint threads
@@ -391,7 +391,7 @@ implements org.apache.hadoop.util.Tool
MultiThreadedClientExample
-publicMultiThreadedClientExample()
+publicMultiThreadedClientExample()
throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
Throws:
@@ -413,7 +413,7 @@ implements org.apache.hadoop.util.Tool
run
-publicintrun(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String[]args)
+publicintrun(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String[]args)
throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
title="class or interface in java.lang">Exception
Specified by:
@@ -429,7 +429,7 @@ implements org.apache.hadoop.util.Tool
warmUpConnectionCache
-privatevoidwarmUpConnectionCache(Connectionconnection,
+privatevoidwarmUpConnectionCache(Connectionconnection,
TableNametn)
throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
@@ -444,7 +444,7 @@ implements org.apache.hadoop.util.Tool
main
-public staticvoidmain(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String[]args)
+public staticvoidmain(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String[]args)
throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
title="class or interface in java.lang">Exception
Throws:
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/071f974b/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
b/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
index 08dd785..2bd2a31 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
@@ -544,24 +544,24 @@
java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableT,
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dc4e5c85/devapidocs/org/apache/hadoop/hbase/executor/EventType.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/executor/EventType.html
b/devapidocs/org/apache/hadoop/hbase/executor/EventType.html
index a57b3c1..fd8332e 100644
--- a/devapidocs/org/apache/hadoop/hbase/executor/EventType.html
+++ b/devapidocs/org/apache/hadoop/hbase/executor/EventType.html
@@ -119,7 +119,7 @@ var activeTableTab = "activeTableTab";
@InterfaceAudience.Private
-public enum EventType
+public enum EventType
extends http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
title="class or interface in java.lang">EnumEventType
List of all HBase event handler types. Event types are
named by a
convention: event type names specify the component from which the event
@@ -513,7 +513,7 @@ the order they are declared.
RS_ZK_REGION_CLOSED
-public static finalEventType RS_ZK_REGION_CLOSED
+public static finalEventType RS_ZK_REGION_CLOSED
RS_ZK_REGION_CLOSED
RS has finished closing a region.
@@ -525,7 +525,7 @@ the order they are declared.
RS_ZK_REGION_OPENING
-public static finalEventType RS_ZK_REGION_OPENING
+public static finalEventType RS_ZK_REGION_OPENING
RS_ZK_REGION_OPENING
RS is in process of opening a region.
@@ -537,7 +537,7 @@ the order they are declared.
RS_ZK_REGION_OPENED
-public static finalEventType RS_ZK_REGION_OPENED
+public static finalEventType RS_ZK_REGION_OPENED
RS_ZK_REGION_OPENED
RS has finished opening a region.
@@ -549,7 +549,7 @@ the order they are declared.
RS_ZK_REGION_SPLITTING
-public static finalEventType RS_ZK_REGION_SPLITTING
+public static finalEventType RS_ZK_REGION_SPLITTING
RS_ZK_REGION_SPLITTING
RS has started a region split after master says it's ok to move on.
@@ -561,7 +561,7 @@ the order they are declared.
RS_ZK_REGION_SPLIT
-public static finalEventType RS_ZK_REGION_SPLIT
+public static finalEventType RS_ZK_REGION_SPLIT
RS_ZK_REGION_SPLIT
RS split has completed and is notifying the master.
@@ -573,7 +573,7 @@ the order they are declared.
RS_ZK_REGION_FAILED_OPEN
-public static finalEventType RS_ZK_REGION_FAILED_OPEN
+public static finalEventType RS_ZK_REGION_FAILED_OPEN
RS_ZK_REGION_FAILED_OPEN
RS failed to open a region.
@@ -585,7 +585,7 @@ the order they are declared.
RS_ZK_REGION_MERGING
-public static finalEventType RS_ZK_REGION_MERGING
+public static finalEventType RS_ZK_REGION_MERGING
RS_ZK_REGION_MERGING
RS has started merging regions after master says it's ok to move on.
@@ -597,7 +597,7 @@ the order they are declared.
RS_ZK_REGION_MERGED
-public static finalEventType RS_ZK_REGION_MERGED
+public static finalEventType RS_ZK_REGION_MERGED
RS_ZK_REGION_MERGE
RS region merge has completed and is notifying the master.
@@ -609,7 +609,7 @@ the order they are declared.
RS_ZK_REQUEST_REGION_SPLIT
-public static finalEventType RS_ZK_REQUEST_REGION_SPLIT
+public static finalEventType RS_ZK_REQUEST_REGION_SPLIT
RS_ZK_REQUEST_REGION_SPLIT
RS has requested to split a region. This is to notify master
@@ -622,7 +622,7 @@ the order they are declared.
RS_ZK_REQUEST_REGION_MERGE
-public static finalEventType RS_ZK_REQUEST_REGION_MERGE
+public static finalEventType RS_ZK_REQUEST_REGION_MERGE
RS_ZK_REQUEST_REGION_MERGE
RS has requested to merge two regions. This is to notify master
@@ -635,7 +635,7 @@ the order they are declared.
M_RS_OPEN_REGION
-public static finalEventType M_RS_OPEN_REGION
+public static finalEventType M_RS_OPEN_REGION
Messages originating from Master to RS.
M_RS_OPEN_REGION
Master asking RS to open a region.
@@ -647,7 +647,7 @@ the order they are declared.
M_RS_OPEN_ROOT
-public static finalEventType M_RS_OPEN_ROOT
+public static finalEventType M_RS_OPEN_ROOT
Messages originating from Master to RS.
M_RS_OPEN_ROOT
Master asking RS to open root.
@@ -659,7 +659,7 @@ the order they are declared.
M_RS_OPEN_META
-public static finalEventType M_RS_OPEN_META
+public static finalEventType M_RS_OPEN_META
Messages originating from Master to RS.
M_RS_OPEN_META
Master asking RS to open meta.
@@ -671,7 +671,7 @@ the order they are declared.
M_RS_CLOSE_REGION
-public static finalEventType M_RS_CLOSE_REGION
+public static finalEventType M_RS_CLOSE_REGION
Messages originating from Master to RS.
M_RS_CLOSE_REGION
Master asking RS to close a region.
@@ -683,7 +683,7 @@ the order they are declared.
M_RS_CLOSE_ROOT
-public static finalEventType M_RS_CLOSE_ROOT
+public static finalEventType M_RS_CLOSE_ROOT
Messages originating from Master to RS.
M_RS_CLOSE_ROOT
Master asking RS to close root.
@@ -695,7 +695,7 @@ the order they are declared.
M_RS_CLOSE_META
-public static finalEventType M_RS_CLOSE_META
+public static finalEventType M_RS_CLOSE_META
Messages originating from Master to RS.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4abd958d/apidocs/src-html/org/apache/hadoop/hbase/HConstants.html
--
diff --git a/apidocs/src-html/org/apache/hadoop/hbase/HConstants.html
b/apidocs/src-html/org/apache/hadoop/hbase/HConstants.html
index 915e78a..abee553 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/HConstants.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/HConstants.html
@@ -36,8 +36,8 @@
028import java.util.regex.Pattern;
029
030import
org.apache.commons.lang3.ArrayUtils;
-031import
org.apache.yetus.audience.InterfaceAudience;
-032import
org.apache.hadoop.hbase.util.Bytes;
+031import
org.apache.hadoop.hbase.util.Bytes;
+032import
org.apache.yetus.audience.InterfaceAudience;
033
034/**
035 * HConstants holds a bunch of
HBase-related constants
@@ -558,786 +558,800 @@
550 /**
551 * Timestamp to use when we want to
refer to the latest cell.
552 *
-553 * On client side, this is the
timestamp set by default when no timestamp is specified, to refer to the
latest.
-554 * On server side, this acts as a
notation.
-555 * (1) For a cell of Put, which has
this notation,
-556 * its timestamp will be replaced
with server's current time.
-557 * (2) For a cell of Delete, which has
this notation,
-558 * A. If the cell is of {@link
KeyValue.Type#Delete}, HBase issues a Get operation firstly.
-559 *a. When the count of cell it
gets is less than the count of cell to delete,
-560 * the timestamp of Delete
cell will be replaced with server's current time.
-561 *b. When the count of cell it
gets is equal to the count of cell to delete,
-562 * the timestamp of Delete
cell will be replaced with the latest timestamp of cell it gets.
-563 * (c. It is invalid and an
exception will be thrown,
-564 * if the count of cell it
gets is greater than the count of cell to delete,
-565 * as the max version of Get
is set to the count of cell to delete.)
-566 * B. If the cell is of other
Delete types, like {@link KeyValue.Type#DeleteFamilyVersion},
-567 *{@link
KeyValue.Type#DeleteColumn}, or {@link KeyValue.Type#DeleteFamily},
-568 *the timestamp of Delete cell
will be replaced with server's current time.
-569 *
-570 * So that is why it is named as
"latest" but assigned as the max value of Long.
-571 */
-572 public static final long
LATEST_TIMESTAMP = Long.MAX_VALUE;
-573
-574 /**
-575 * Timestamp to use when we want to
refer to the oldest cell.
-576 * Special! Used in fake Cells only.
Should never be the timestamp on an actual Cell returned to
-577 * a client.
-578 * @deprecated Should not be public
since hbase-1.3.0. For internal use only. Move internal to
-579 * Scanners flagged as special
timestamp value never to be returned as timestamp on a Cell.
-580 */
-581 @Deprecated
-582 public static final long
OLDEST_TIMESTAMP = Long.MIN_VALUE;
-583
-584 /**
-585 * LATEST_TIMESTAMP in bytes form
-586 */
-587 public static final byte []
LATEST_TIMESTAMP_BYTES = {
-588// big-endian
-589(byte) (LATEST_TIMESTAMP
56),
-590(byte) (LATEST_TIMESTAMP
48),
-591(byte) (LATEST_TIMESTAMP
40),
-592(byte) (LATEST_TIMESTAMP
32),
-593(byte) (LATEST_TIMESTAMP
24),
-594(byte) (LATEST_TIMESTAMP
16),
-595(byte) (LATEST_TIMESTAMP
8),
-596(byte) LATEST_TIMESTAMP,
-597 };
-598
-599 /**
-600 * Define for 'return-all-versions'.
-601 */
-602 public static final int ALL_VERSIONS =
Integer.MAX_VALUE;
-603
-604 /**
-605 * Unlimited time-to-live.
-606 */
-607// public static final int FOREVER =
-1;
-608 public static final int FOREVER =
Integer.MAX_VALUE;
-609
-610 /**
-611 * Seconds in a day, hour and minute
-612 */
-613 public static final int DAY_IN_SECONDS
= 24 * 60 * 60;
-614 public static final int HOUR_IN_SECONDS
= 60 * 60;
-615 public static final int
MINUTE_IN_SECONDS = 60;
-616
-617 //TODO: although the following are
referenced widely to format strings for
-618 // the shell. They really aren't a
part of the public API. It would be
-619 // nice if we could put them
somewhere where they did not need to be
-620 // public. They could have package
visibility
-621 public static final String NAME =
"NAME";
-622 public static final String VERSIONS =
"VERSIONS";
-623 public static final String IN_MEMORY =
"IN_MEMORY";
-624 public static final String METADATA =
"METADATA";
-625 public static final String
CONFIGURATION = "CONFIGURATION";
-626
-627 /**
-628 * Retrying we multiply
hbase.client.pause setting by what we have in this array until we
-629 * run out of array items. Retries
beyond this use the last number in the array. So, for
-630 * example, if hbase.client.pause is 1
second, and maximum retries count
-631 * hbase.client.retries.number is 10,
we will
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e23b49ba/apidocs/src-html/org/apache/hadoop/hbase/ClusterStatus.html
--
diff --git a/apidocs/src-html/org/apache/hadoop/hbase/ClusterStatus.html
b/apidocs/src-html/org/apache/hadoop/hbase/ClusterStatus.html
index 1aaa281..c3eca54 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/ClusterStatus.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/ClusterStatus.html
@@ -201,297 +201,303 @@
193
194 /**
195 * @return the number of requests since
last report
-196 */
-197 public int getRequestsCount() {
-198int count = 0;
-199if (liveServers != null
!liveServers.isEmpty()) {
-200 for (Map.EntryServerName,
ServerLoad e: this.liveServers.entrySet()) {
-201count +=
e.getValue().getNumberOfRequests();
-202 }
-203}
-204return count;
-205 }
-206
-207 /**
-208 * @return the HBase version string as
reported by the HMaster
-209 */
-210 public String getHBaseVersion() {
-211return hbaseVersion;
-212 }
-213
-214 /**
-215 * @see
java.lang.Object#equals(java.lang.Object)
-216 */
-217 public boolean equals(Object o) {
-218if (this == o) {
-219 return true;
-220}
-221if (!(o instanceof ClusterStatus))
{
-222 return false;
-223}
-224ClusterStatus other = (ClusterStatus)
o;
-225return
Objects.equal(getHBaseVersion(), other.getHBaseVersion())
-226 Objects.equal(this.liveServers,
other.liveServers)
-227
getDeadServerNames().containsAll(other.getDeadServerNames())
-228
Arrays.equals(getMasterCoprocessors(), other.getMasterCoprocessors())
-229 Objects.equal(getMaster(),
other.getMaster())
-230
getBackupMasters().containsAll(other.getBackupMasters())
-231 Objects.equal(getClusterId(),
other.getClusterId())
-232 getMasterInfoPort() ==
other.getMasterInfoPort();
-233 }
-234
-235 /**
-236 * @see java.lang.Object#hashCode()
-237 */
-238 public int hashCode() {
-239return Objects.hashCode(hbaseVersion,
liveServers, deadServers, master, backupMasters,
-240 clusterId, masterInfoPort);
-241 }
-242
-243 /**
-244 *
-245 * @return the object version number
-246 * @deprecated As of release 2.0.0,
this will be removed in HBase 3.0.0
-247 */
-248 @Deprecated
-249 public byte getVersion() {
-250return VERSION;
-251 }
-252
-253 //
-254 // Getters
-255 //
-256
-257 public CollectionServerName
getServers() {
-258if (liveServers == null) {
-259 return Collections.emptyList();
-260}
-261return
Collections.unmodifiableCollection(this.liveServers.keySet());
-262 }
-263
-264 /**
-265 * Returns detailed information about
the current master {@link ServerName}.
-266 * @return current master information
if it exists
-267 */
-268 public ServerName getMaster() {
-269return this.master;
-270 }
-271
-272 /**
-273 * @return the number of backup masters
in the cluster
-274 */
-275 public int getBackupMastersSize() {
-276return backupMasters != null ?
backupMasters.size() : 0;
-277 }
-278
-279 /**
-280 * @return the names of backup
masters
-281 */
-282 public ListServerName
getBackupMasters() {
-283if (backupMasters == null) {
-284 return Collections.emptyList();
-285}
-286return
Collections.unmodifiableList(this.backupMasters);
-287 }
-288
-289 /**
-290 * @param sn
-291 * @return Server's load or null if not
found.
-292 */
-293 public ServerLoad getLoad(final
ServerName sn) {
-294return liveServers != null ?
liveServers.get(sn) : null;
-295 }
-296
-297 @InterfaceAudience.Private
-298 public ListRegionState
getRegionsInTransition() {
-299if (intransition == null) {
-300 return Collections.emptyList();
-301}
-302return
Collections.unmodifiableList(intransition);
-303 }
-304
-305 public String getClusterId() {
-306return clusterId;
-307 }
-308
-309 public String[] getMasterCoprocessors()
{
-310return masterCoprocessors;
-311 }
-312
-313 public long
getLastMajorCompactionTsForTable(TableName table) {
-314long result = Long.MAX_VALUE;
-315for (ServerName server :
getServers()) {
-316 ServerLoad load =
getLoad(server);
-317 for (RegionLoad rl :
load.getRegionsLoad().values()) {
-318if
(table.equals(HRegionInfo.getTable(rl.getName( {
-319 result = Math.min(result,
rl.getLastMajorCompactionTs());
-320}
-321 }
-322}
-323return result == Long.MAX_VALUE ? 0 :
result;
-324 }
-325
-326 public long
getLastMajorCompactionTsForRegion(final byte[] region) {
-327for (ServerName server :
getServers()) {
-328 ServerLoad load =
getLoad(server);
-329 RegionLoad rl =
load.getRegionsLoad().get(region);
-330 if (rl != null) {
-331return
rl.getLastMajorCompactionTs();
-332 }
-333}
-334return 0;
-335 }
-336
-337 public boolean
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html
--
diff --git
a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html
b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html
index 62bc799..5c004ce 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.html
@@ -250,7 +250,7 @@
242Cell kv = cell;
243// null input == user explicitly
wants to flush
244if (row == null kv ==
null) {
-245 rollWriters();
+245 rollWriters(null);
246 return;
247}
248
@@ -284,636 +284,642 @@
276 configureStoragePolicy(conf,
fs, tableAndFamily, writerPath);
277}
278
-279// If any of the HFiles for the
column families has reached
-280// maxsize, we need to roll all
the writers
-281if (wl != null
wl.written + length = maxsize) {
-282 this.rollRequested = true;
-283}
-284
-285// This can only happen once a
row is finished though
-286if (rollRequested
Bytes.compareTo(this.previousRow, rowKey) != 0) {
-287 rollWriters();
-288}
-289
-290// create a new WAL writer, if
necessary
-291if (wl == null || wl.writer ==
null) {
-292 if
(conf.getBoolean(LOCALITY_SENSITIVE_CONF_KEY, DEFAULT_LOCALITY_SENSITIVE)) {
-293HRegionLocation loc = null;
-294
-295String tableName =
Bytes.toString(tableNameBytes);
-296if (tableName != null) {
-297 try (Connection connection
= ConnectionFactory.createConnection(conf);
-298 RegionLocator
locator =
-299
connection.getRegionLocator(TableName.valueOf(tableName))) {
-300loc =
locator.getRegionLocation(rowKey);
-301 } catch (Throwable e) {
-302LOG.warn("There's
something wrong when locating rowkey: " +
-303 Bytes.toString(rowKey)
+ " for tablename: " + tableName, e);
-304loc = null;
-305 } }
-306
-307if (null == loc) {
-308 if (LOG.isTraceEnabled())
{
-309LOG.trace("failed to get
region location, so use default writer for rowkey: " +
-310
Bytes.toString(rowKey));
-311 }
-312 wl =
getNewWriter(tableNameBytes, family, conf, null);
-313} else {
-314 if (LOG.isDebugEnabled())
{
-315LOG.debug("first rowkey:
[" + Bytes.toString(rowKey) + "]");
-316 }
-317 InetSocketAddress
initialIsa =
-318 new
InetSocketAddress(loc.getHostname(), loc.getPort());
-319 if
(initialIsa.isUnresolved()) {
-320if (LOG.isTraceEnabled())
{
-321 LOG.trace("failed to
resolve bind address: " + loc.getHostname() + ":"
-322 + loc.getPort() +
", so use default writer");
-323}
-324wl =
getNewWriter(tableNameBytes, family, conf, null);
-325 } else {
-326if (LOG.isDebugEnabled())
{
-327 LOG.debug("use favored
nodes writer: " + initialIsa.getHostString());
-328}
-329wl =
getNewWriter(tableNameBytes, family, conf, new InetSocketAddress[] {
initialIsa
-330});
-331 }
-332}
-333 } else {
-334wl =
getNewWriter(tableNameBytes, family, conf, null);
-335 }
-336}
-337
-338// we now have the proper WAL
writer. full steam ahead
-339// TODO : Currently in
SettableTimeStamp but this will also move to ExtendedCell
-340
PrivateCellUtil.updateLatestStamp(cell, this.now);
-341wl.writer.append(kv);
-342wl.written += length;
-343
-344// Copy the row so we know when a
row transition.
-345this.previousRow = rowKey;
-346 }
-347
-348 private void rollWriters() throws
IOException {
-349for (WriterLength wl :
this.writers.values()) {
-350 if (wl.writer != null) {
-351LOG.info(
-352"Writer=" +
wl.writer.getPath() + ((wl.written == 0)? "": ", wrote=" + wl.written));
-353close(wl.writer);
-354 }
-355 wl.writer = null;
-356 wl.written = 0;
-357}
-358this.rollRequested = false;
-359 }
-360
-361 /*
-362 * Create a new StoreFile.Writer.
-363 * @param family
-364 * @return A WriterLength,
containing a new StoreFile.Writer.
-365 * @throws IOException
-366 */
-367
@edu.umd.cs.findbugs.annotations.SuppressWarnings(value="BX_UNBOXING_IMMEDIATELY_REBOXED",
-368
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c54c242b/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
b/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
index e1de268..c93055b 100644
---
a/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
+++
b/devapidocs/org/apache/hadoop/hbase/regionserver/HStore.StoreFlusherImpl.html
@@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
-private final class HStore.StoreFlusherImpl
+private final class HStore.StoreFlusherImpl
extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
implements StoreFlushContext
@@ -238,7 +238,7 @@ implements getOutputFileSize()
-void
+MemStoreSize
prepare()
This is not thread safe.
@@ -279,7 +279,7 @@ implements
tracker
-private finalFlushLifeCycleTracker tracker
+private finalFlushLifeCycleTracker tracker
@@ -288,7 +288,7 @@ implements
cacheFlushSeqNum
-private finallong cacheFlushSeqNum
+private finallong cacheFlushSeqNum
@@ -297,7 +297,7 @@ implements
snapshot
-privateMemStoreSnapshot snapshot
+privateMemStoreSnapshot snapshot
@@ -306,7 +306,7 @@ implements
tempFiles
-privatehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in
java.util">Listorg.apache.hadoop.fs.Path tempFiles
+privatehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in
java.util">Listorg.apache.hadoop.fs.Path tempFiles
@@ -315,7 +315,7 @@ implements
committedFiles
-privatehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in
java.util">Listorg.apache.hadoop.fs.Path committedFiles
+privatehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in
java.util">Listorg.apache.hadoop.fs.Path committedFiles
@@ -324,7 +324,7 @@ implements
cacheFlushCount
-privatelong cacheFlushCount
+privatelong cacheFlushCount
@@ -333,7 +333,7 @@ implements
cacheFlushSize
-privatelong cacheFlushSize
+privatelong cacheFlushSize
@@ -342,7 +342,7 @@ implements
outputFileSize
-privatelong outputFileSize
+privatelong outputFileSize
@@ -359,7 +359,7 @@ implements
StoreFlusherImpl
-privateStoreFlusherImpl(longcacheFlushSeqNum,
+privateStoreFlusherImpl(longcacheFlushSeqNum,
FlushLifeCycleTrackertracker)
@@ -377,12 +377,14 @@ implements
prepare
-publicvoidprepare()
+publicMemStoreSizeprepare()
This is not thread safe. The caller should have a lock on
the region or the store.
If necessary, the lock can be added with the patch provided in
HBASE-10087
Specified by:
preparein
interfaceStoreFlushContext
+Returns:
+The size of snapshot to flush
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713d773f/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
b/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
index ae84eb3..3480c51 100644
--- a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
@@ -168,9 +168,9 @@
java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
title="class or interface in java.io">Serializable)
org.apache.hadoop.hbase.backup.BackupInfo.BackupPhase
+org.apache.hadoop.hbase.backup.BackupInfo.BackupState
org.apache.hadoop.hbase.backup.BackupType
org.apache.hadoop.hbase.backup.BackupRestoreConstants.BackupCommand
-org.apache.hadoop.hbase.backup.BackupInfo.BackupState
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713d773f/devapidocs/org/apache/hadoop/hbase/class-use/ServerName.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/ServerName.html
b/devapidocs/org/apache/hadoop/hbase/class-use/ServerName.html
index f6741a4..a1176d2 100644
--- a/devapidocs/org/apache/hadoop/hbase/class-use/ServerName.html
+++ b/devapidocs/org/apache/hadoop/hbase/class-use/ServerName.html
@@ -3975,11 +3975,6 @@ Input/OutputFormats, a table indexing MapReduce job, and
utility methods.
-RegionStates.ServerStateNode
-RegionStates.addRegionToServer(ServerNameserverName,
- RegionStates.RegionStateNoderegionNode)
-
-
protected boolean
RegionTransitionProcedure.addToRemoteDispatcher(MasterProcedureEnvenv,
ServerNametargetServer)
@@ -3989,31 +3984,31 @@ Input/OutputFormats, a table indexing MapReduce job,
and utility methods.
to pick it up.
-
+
void
AssignmentManager.assignMeta(RegionInfometaRegionInfo,
ServerNameserverName)
-
+
AssignProcedure
AssignmentManager.createAssignProcedure(RegionInforegionInfo,
ServerNametargetServer)
-
+
UnassignProcedure
AssignmentManager.createUnassignProcedure(RegionInforegionInfo,
ServerNamedestinationServer,
booleanforce)
-
+
private http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListRegionInfo
AssignmentManager.getCarryingSystemTables(ServerNameserverName)
-
+
RegionStates.ServerStateNode
RegionStates.getOrCreateServer(ServerNameserverName)
-
+
(package private) static
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse
Util.getRegionInfoResponse(MasterProcedureEnvenv,
ServerNameregionLocation,
@@ -4021,50 +4016,50 @@ Input/OutputFormats, a table indexing MapReduce job,
and utility methods.
Raw call to remote regionserver to get info on a particular
region.
-
+
(package private) static
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse
Util.getRegionInfoResponse(MasterProcedureEnvenv,
ServerNameregionLocation,
RegionInfohri,
booleanincludeBestSplitRow)
-
+
protected RegionStates.ServerStateNode
RegionStates.getServerNode(ServerNameserverName)
-
+
http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListRegionInfo
RegionStates.getServerRegionInfoSet(ServerNameserverName)
Returns the set of regions hosted by the specified
server
-
+
int
AssignmentManager.getServerVersion(ServerNameserverName)
-
+
boolean
AssignmentManager.isCarryingMeta(ServerNameserverName)
-
+
private boolean
AssignmentManager.isCarryingRegion(ServerNameserverName,
RegionInforegionInfo)
-
+
protected boolean
RegionTransitionProcedure.isServerOnline(MasterProcedureEnvenv,
ServerNameserverName)
-
+
void
AssignmentManager.killRegionServer(ServerNameserverName)
-
+
void
RegionStates.logSplit(ServerNameserverName)
-
+
void
AssignmentManager.markRegionAsMerged(RegionInfochild,
ServerNameserverName,
@@ -4073,62 +4068,62 @@ Input/OutputFormats, a table indexing MapReduce job,
and utility methods.
When called here, the merge has happened.
-
+
void
AssignmentManager.markRegionAsSplit(RegionInfoparent,
ServerNameserverName,
RegionInfodaughterA,
RegionInfodaughterB)
-
+
void
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fd365a2b/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html
--
diff --git
a/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html
b/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html
index 5e56349..92bdee3 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html
@@ -259,680 +259,689 @@
251 }
252
253 @Override
-254 public HTableDescriptor
getTableDescriptor() throws IOException {
-255StringBuilder sb = new
StringBuilder();
-256sb.append('/');
-257sb.append(Bytes.toString(name));
-258sb.append('/');
-259sb.append("schema");
-260for (int i = 0; i maxRetries;
i++) {
-261 Response response =
client.get(sb.toString(), Constants.MIMETYPE_PROTOBUF);
-262 int code = response.getCode();
-263 switch (code) {
-264 case 200:
-265TableSchemaModel schema = new
TableSchemaModel();
-266
schema.getObjectFromMessage(response.getBody());
-267return
schema.getTableDescriptor();
-268 case 509:
-269try {
-270 Thread.sleep(sleepTime);
-271} catch (InterruptedException e)
{
-272 throw
(InterruptedIOException)new InterruptedIOException().initCause(e);
-273}
-274break;
-275 default:
-276throw new IOException("schema
request returned " + code);
-277 }
-278}
-279throw new IOException("schema request
timed out");
-280 }
-281
-282 @Override
-283 public void close() throws IOException
{
-284client.shutdown();
-285 }
-286
-287 @Override
-288 public Result get(Get get) throws
IOException {
-289TimeRange range =
get.getTimeRange();
-290String spec =
buildRowSpec(get.getRow(), get.getFamilyMap(),
-291 range.getMin(), range.getMax(),
get.getMaxVersions());
-292if (get.getFilter() != null) {
-293 LOG.warn("filters not supported on
gets");
-294}
-295Result[] results =
getResults(spec);
-296if (results.length 0) {
-297 if (results.length 1) {
-298LOG.warn("too many results for
get (" + results.length + ")");
-299 }
-300 return results[0];
-301} else {
-302 return new Result();
-303}
-304 }
-305
-306 @Override
-307 public Result[] get(ListGet
gets) throws IOException {
-308byte[][] rows = new
byte[gets.size()][];
-309int maxVersions = 1;
-310int count = 0;
-311
-312for(Get g:gets) {
-313
-314 if ( count == 0 ) {
-315maxVersions =
g.getMaxVersions();
-316 } else if (g.getMaxVersions() !=
maxVersions) {
-317LOG.warn("MaxVersions on Gets do
not match, using the first in the list ("+maxVersions+")");
-318 }
-319
-320 if (g.getFilter() != null) {
-321LOG.warn("filters not supported
on gets");
-322 }
-323
-324 rows[count] = g.getRow();
-325 count ++;
-326}
-327
-328String spec = buildMultiRowSpec(rows,
maxVersions);
-329
-330return getResults(spec);
-331 }
-332
-333 private Result[] getResults(String
spec) throws IOException {
-334for (int i = 0; i maxRetries;
i++) {
-335 Response response =
client.get(spec, Constants.MIMETYPE_PROTOBUF);
-336 int code = response.getCode();
-337 switch (code) {
-338case 200:
-339 CellSetModel model = new
CellSetModel();
-340
model.getObjectFromMessage(response.getBody());
-341 Result[] results =
buildResultFromModel(model);
-342 if ( results.length 0) {
-343return results;
-344 }
-345 // fall through
-346case 404:
-347 return new Result[0];
-348
-349case 509:
-350 try {
-351Thread.sleep(sleepTime);
-352 } catch (InterruptedException
e) {
-353throw
(InterruptedIOException)new InterruptedIOException().initCause(e);
-354 }
-355 break;
-356default:
-357 throw new IOException("get
request returned " + code);
-358 }
-359}
-360throw new IOException("get request
timed out");
-361 }
-362
-363 @Override
-364 public boolean exists(Get get) throws
IOException {
-365LOG.warn("exists() is really get(),
just use get()");
-366Result result = get(get);
-367return (result != null
!(result.isEmpty()));
-368 }
-369
-370 @Override
-371 public boolean[] exists(ListGet
gets) throws IOException {
-372LOG.warn("exists(ListGet) is
really list of get() calls, just use get()");
-373boolean[] results = new
boolean[gets.size()];
-374for (int i = 0; i
results.length; i++) {
-375 results[i] = exists(gets.get(i));
-376}
-377return results;
-378 }
-379
-380 @Override
-381 public void put(Put put) throws
IOException {
-382CellSetModel model =
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b9722a17/devapidocs/org/apache/hadoop/hbase/procedure2/AbstractProcedureScheduler.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/procedure2/AbstractProcedureScheduler.html
b/devapidocs/org/apache/hadoop/hbase/procedure2/AbstractProcedureScheduler.html
index ce260ce..435c524 100644
---
a/devapidocs/org/apache/hadoop/hbase/procedure2/AbstractProcedureScheduler.html
+++
b/devapidocs/org/apache/hadoop/hbase/procedure2/AbstractProcedureScheduler.html
@@ -18,7 +18,7 @@
catch(err) {
}
//-->
-var methods =
{"i0":10,"i1":10,"i2":6,"i3":6,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":6,"i12":6,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10};
+var methods =
{"i0":10,"i1":10,"i2":10,"i3":6,"i4":6,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":6,"i13":6,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10};
var tabs = {65535:["t0","All Methods"],2:["t2","Instance
Methods"],4:["t3","Abstract Methods"],8:["t4","Concrete Methods"]};
var altColor = "altColor";
var rowColor = "rowColor";
@@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab";
@InterfaceAudience.Private
-public abstract class AbstractProcedureScheduler
+public abstract class AbstractProcedureScheduler
extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
implements ProcedureScheduler
@@ -203,145 +203,130 @@ implements
void
+addFront(http://docs.oracle.com/javase/8/docs/api/java/util/Iterator.html?is-external=true;
title="class or interface in java.util">IteratorProcedureprocedureIterator)
+Inserts all elements in the iterator at the front of this
queue.
+
+
+
+void
addFront(Procedureprocedure)
Inserts the specified element at the front of this
queue.
-
+
protected abstract Procedure
dequeue()
Fetch one Procedure from the queue
NOTE: this method is called with the sched lock held.
-
+
protected abstract void
enqueue(Procedureprocedure,
booleanaddFront)
Add the procedure to the queue.
-
+
long
getNullPollCalls()
-
+
long
getPollCalls()
-
+
boolean
hasRunnables()
-
+
Procedure
poll()
Fetch one Procedure from the queue
-
+
Procedure
poll(longnanos)
-
+
Procedure
poll(longtimeout,
http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true;
title="class or interface in
java.util.concurrent">TimeUnitunit)
Fetch one Procedure from the queue
-
+
protected void
push(Procedureprocedure,
booleanaddFront,
booleannotify)
-
+
protected abstract boolean
queueHasRunnables()
Returns true if there are procedures available to
process.
-
+
protected abstract int
queueSize()
Returns the number of elements in this queue.
-
+
protected void
schedLock()
-
+
protected void
schedUnlock()
-
+
void
signalAll()
In case the class is blocking on poll() waiting for items
to be added,
this method should awake poll() and poll() should return.
-
+
int
size()
Returns the number of elements in this queue.
-
+
void
start()
Start the scheduler
-
+
void
stop()
Stop the scheduler
-
-void
-suspendEvent(ProcedureEventevent)
-Mark the event as not ready.
-
-
-boolean
-waitEvent(ProcedureEventevent,
- Procedureprocedure)
-Suspend the procedure if the event is not ready yet.
-
-
-
protected void
waitProcedure(ProcedureDequewaitQueue,
Procedureproc)
-
-void
-wakeEvent(ProcedureEventevent)
-Wake every procedure waiting for the specified event
- (By design each event has only one "wake" caller)
-
-
-
-void
-wakeEvents(intcount,
- ProcedureEvent...events)
-Wake every procedure waiting for the specified events.
+
+(package private) void
+wakeEvents(ProcedureEvent[]events)
+Wake up all of the given events.
-
+
protected void
wakePollIfNeeded(intwaitingCount)
-
+
protected void
wakeProcedure(Procedureprocedure)
-
+
protected int
wakeWaitingProcedures(ProcedureDequewaitQueue)
Wakes up given waiting procedures by pushing them back into
scheduler queues.
@@ -382,7 +367,7 @@ implements
LOG
-private static finalorg.apache.commons.logging.Log LOG
+private static finalorg.apache.commons.logging.Log LOG
@@ -391,7 +376,7 @@ implements
schedulerLock
-private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/ReentrantLock.html?is-external=true;
title="class or interface in java.util.concurrent.locks">ReentrantLock schedulerLock
+private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/ReentrantLock.html?is-external=true;
title="class or interface in java.util.concurrent.locks">ReentrantLock schedulerLock
@@ -400,7 +385,7 @@
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b1eb7453/devapidocs/deprecated-list.html
--
diff --git a/devapidocs/deprecated-list.html b/devapidocs/deprecated-list.html
index 80ca7f4..ca12b98 100644
--- a/devapidocs/deprecated-list.html
+++ b/devapidocs/deprecated-list.html
@@ -523,7 +523,8 @@
org.apache.hadoop.hbase.CellUtil.cloneTags(Cell)
-As of HBase-2.0. Will be
removed in HBase-3.0.
+As of HBase-2.0. Will be
removed in HBase-3.0.
+ Use RawCell.cloneTags()
@@ -687,67 +688,85 @@
+org.apache.hadoop.hbase.CellUtil.createCell(Cell,
byte[])
+As of release 2.0.0, this
will be removed in HBase 3.0.0.
+Use CP environment to build Cell using ExtendedCellBuilder
+
+
+
+org.apache.hadoop.hbase.CellUtil.createCell(Cell,
byte[], byte[])
+As of release 2.0.0, this
will be removed in HBase 3.0.0.
+ Use CP environment to build Cell using ExtendedCellBuilder
+
+
+
+org.apache.hadoop.hbase.CellUtil.createCell(Cell,
ListTag)
+As of release 2.0.0, this
will be removed in HBase 3.0.0.
+ Use CP environment to build Cell using ExtendedCellBuilder
+
+
+
org.apache.hadoop.hbase.filter.ParseFilter.createCompareOp(byte[])
Since 2.0
-
+
org.apache.hadoop.hbase.filter.ParseFilter.createCompareOperator(byte[])
Since 2.0
-
+
org.apache.hadoop.hbase.HRegionInfo.createRegionName(TableName,
byte[], byte[], boolean)
As of release 2.0.0, this
will be removed in HBase 3.0.0
Use RegionInfo.createRegionName(TableName,
byte[], byte[], boolean).
-
+
org.apache.hadoop.hbase.HRegionInfo.createRegionName(TableName,
byte[], byte[], int, boolean)
As of release 2.0.0, this
will be removed in HBase 3.0.0
Use RegionInfo.createRegionName(TableName,
byte[], byte[], int, boolean).
-
+
org.apache.hadoop.hbase.HRegionInfo.createRegionName(TableName,
byte[], long, boolean)
As of release 2.0.0, this
will be removed in HBase 3.0.0
Use RegionInfo.createRegionName(TableName,
byte[], long, boolean).
-
+
org.apache.hadoop.hbase.HRegionInfo.createRegionName(TableName,
byte[], long, int, boolean)
As of release 2.0.0, this
will be removed in HBase 3.0.0
Use RegionInfo.createRegionName(TableName,
byte[], long, int, boolean).
-
+
org.apache.hadoop.hbase.HRegionInfo.createRegionName(TableName,
byte[], String, boolean)
As of release 2.0.0, this
will be removed in HBase 3.0.0
Use RegionInfo.createRegionName(TableName,
byte[], String, boolean).
-
+
org.apache.hadoop.hbase.client.Admin.deleteColumn(TableName,
byte[])
As of release 2.0.0.
This will be removed in HBase 3.0.0.
Use Admin.deleteColumnFamily(TableName,
byte[])}.
-
+
org.apache.hadoop.hbase.client.HBaseAdmin.deleteColumn(TableName,
byte[])
Since 2.0. Will be removed
in 3.0. Use
HBaseAdmin.deleteColumnFamily(TableName,
byte[]) instead.
-
+
org.apache.hadoop.hbase.client.Admin.deleteSnapshots(String)
since 2.0 version and will
be removed in 3.0 version.
Use Admin.deleteSnapshots(Pattern)
instead.
-
+
org.apache.hadoop.hbase.client.Admin.deleteTables(Pattern)
since 2.0 version and will
be removed in 3.0 version
This is just a trivial helper method without any magic.
@@ -755,7 +774,7 @@
and Admin.deleteTable(TableName)
-
+
org.apache.hadoop.hbase.client.Admin.deleteTables(String)
since 2.0 version and will
be removed in 3.0 version
This is just a trivial helper method without any magic.
@@ -763,25 +782,25 @@
and Admin.deleteTable(TableName)
-
+
org.apache.hadoop.hbase.client.Admin.deleteTableSnapshots(String,
String)
since 2.0 version and will
be removed in 3.0 version.
Use Admin.deleteTableSnapshots(Pattern,
Pattern) instead.
-
+
org.apache.hadoop.hbase.client.replication.ReplicationAdmin.disablePeer(String)
use Admin.disableReplicationPeer(String)
instead
-
+
org.apache.hadoop.hbase.client.replication.ReplicationAdmin.disableTableRep(TableName)
use Admin.disableTableReplication(TableName)
instead
-
+
org.apache.hadoop.hbase.client.Admin.disableTables(Pattern)
since 2.0 version and will
be removed in 3.0 version
This is just a trivial helper method without any magic.
@@ -789,7 +808,7 @@
and Admin.disableTable(org.apache.hadoop.hbase.TableName)
-
+
org.apache.hadoop.hbase.client.Admin.disableTables(String)
since 2.0 version and will
be removed in 3.0 version
This is just a trivial helper method without any magic.
@@ -797,25 +816,25 @@
and Admin.disableTable(org.apache.hadoop.hbase.TableName)
-
+
org.apache.hadoop.hbase.client.Admin.enableCatalogJanitor(boolean)
Since 2.0.0. Will be
removed in 3.0.0. Use Admin.catalogJanitorSwitch(boolean)}
instead.
-
+
1 - 100 of 294 matches
Mail list logo
|