http://git-wip-us.apache.org/repos/asf/hbase-site/blob/5427a45e/apidocs/org/apache/hadoop/hbase/class-use/CellScanner.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/class-use/CellScanner.html
b/apidocs/org/apache/hadoop/hbase/class-use/CellScanner.html
index 6b2e30c..dae577c 100644
--- a/apidocs/org/apache/hadoop/hbase/class-use/CellScanner.html
+++ b/apidocs/org/apache/hadoop/hbase/class-use/CellScanner.html
@@ -1,10 +1,10 @@
http://www.w3.org/TR/html4/loose.dtd;>
-
+
-Uses of Interface org.apache.hadoop.hbase.CellScanner (Apache HBase
3.0.0-SNAPSHOT API)
+æ¥å£ org.apache.hadoop.hbase.CellScannerçä½¿ç¨ (Apache HBase
3.0.0-SNAPSHOT API)
@@ -12,7 +12,7 @@
-JavaScript is disabled on your browser.
+æ¨çæµè§å¨å·²ç¦ç¨ JavaScriptã
-Skip navigation links
+è·³è¿å¯¼èªé¾æ¥
-
-Overview
-Package
-Class
-Use
-Tree
-Deprecated
-Index
-Help
+
+æ¦è§
+ç¨åºå
+ç±»
+使ç¨
+æ
+å·²è¿æ¶
+ç´¢å¼
+帮å©
-Prev
-Next
+ä¸ä¸ä¸ª
+ä¸ä¸ä¸ª
-Frames
-NoFrames
+æ¡æ¶
+æ æ¡æ¶
-AllClasses
+ææç±»
-
-
-Packages that use CellScanner
+
+使ç¨CellScannerçç¨åºå
-Package |
-Description |
+ç¨åºå
|
+说æ |
@@ -89,7 +89,17 @@
org.apache.hadoop.hbase.client |
- Provides HBase Client
+
|
@@ -100,37 +110,37 @@
-
-
-
-Methods in org.apache.hadoop.hbase
that return CellScanner
+
+
+è¿å
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bcb555af/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/BlockCacheUtil.CachedBlockCountsPerFile.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/BlockCacheUtil.CachedBlockCountsPerFile.html
b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/BlockCacheUtil.CachedBlockCountsPerFile.html
index f118c08..cd9722f 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/BlockCacheUtil.CachedBlockCountsPerFile.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/BlockCacheUtil.CachedBlockCountsPerFile.html
@@ -206,140 +206,175 @@
198 */
199 public static int
validateBlockAddition(Cacheable existing, Cacheable newBlock,
200
BlockCacheKey cacheKey) {
-201int comparison =
compareCacheBlock(existing, newBlock, true);
+201int comparison =
compareCacheBlock(existing, newBlock, false);
202if (comparison != 0) {
-203 LOG.warn("Cached block contents
differ, trying to just compare the block contents " +
-204 "without the next block.
CacheKey: " + cacheKey);
-205
-206 // compare the contents, if they
are not equal, we are in big trouble
-207 int
comparisonWithoutNextBlockMetadata = compareCacheBlock(existing, newBlock,
false);
-208
-209 if
(comparisonWithoutNextBlockMetadata != 0) {
-210throw new
RuntimeException("Cached block contents differ, which should not have
happened."
-211+ "cacheKey:" + cacheKey);
-212 }
-213}
-214return comparison;
-215 }
-216
-217 /**
-218 * Use one of these to keep a running
account of cached blocks by file. Throw it away when done.
-219 * This is different than metrics in
that it is stats on current state of a cache.
-220 * See getLoadedCachedBlocksByFile
-221 */
-222
@JsonIgnoreProperties({"cachedBlockStatsByFile"})
-223 public static class CachedBlocksByFile
{
-224private int count;
-225private int dataBlockCount;
-226private long size;
-227private long dataSize;
-228private final long now =
System.nanoTime();
-229/**
-230 * How many blocks to look at before
we give up.
-231 * There could be many millions of
blocks. We don't want the
-232 * ui to freeze while we run through
1B blocks... users will
-233 * think hbase dead. UI displays
warning in red when stats
-234 * are incomplete.
-235 */
-236private final int max;
-237public static final int DEFAULT_MAX =
100;
-238
-239CachedBlocksByFile() {
-240 this(null);
-241}
-242
-243CachedBlocksByFile(final
Configuration c) {
-244 this.max = c == null? DEFAULT_MAX:
c.getInt("hbase.ui.blockcache.by.file.max", DEFAULT_MAX);
-245}
-246
-247/**
-248 * Map by filename. use concurent
utils because we want our Map and contained blocks sorted.
-249 */
-250private NavigableMapString,
NavigableSetCachedBlock cachedBlockByFile = new
ConcurrentSkipListMap();
-251FastLongHistogram hist = new
FastLongHistogram();
-252
-253/**
-254 * @param cb
-255 * @return True if full if we
won't be adding any more.
-256 */
-257public boolean update(final
CachedBlock cb) {
-258 if (isFull()) return true;
-259 NavigableSetCachedBlock set
= this.cachedBlockByFile.get(cb.getFilename());
-260 if (set == null) {
-261set = new
ConcurrentSkipListSet();
-262
this.cachedBlockByFile.put(cb.getFilename(), set);
-263 }
-264 set.add(cb);
-265 this.size += cb.getSize();
-266 this.count++;
-267 BlockType bt = cb.getBlockType();
-268 if (bt != null
bt.isData()) {
-269this.dataBlockCount++;
-270this.dataSize += cb.getSize();
-271 }
-272 long age = (this.now -
cb.getCachedTime())/NANOS_PER_SECOND;
-273 this.hist.add(age, 1);
-274 return false;
-275}
-276
-277/**
-278 * @return True if full; i.e. there
are more items in the cache but we only loaded up
-279 * the maximum set in configuration
codehbase.ui.blockcache.by.file.max/code
-280 * (Default: DEFAULT_MAX).
-281 */
-282public boolean isFull() {
-283 return this.count = this.max;
-284}
-285
-286public NavigableMapString,
NavigableSetCachedBlock getCachedBlockStatsByFile() {
-287 return this.cachedBlockByFile;
-288}
-289
-290/**
-291 * @return count of blocks in the
cache
-292 */
-293public int getCount() {
-294 return count;
-295}
-296
-297public int getDataCount() {
-298 return dataBlockCount;
-299}
-300
-301/**
-302 * @return size of blocks in the
cache
-303 */
-304public long getSize() {
-305 return size;
-306}
-307
-308/**
-309 * @return Size of data.
-310 */
-311public long getDataSize() {
-312 return dataSize;
-313}
-314
-315
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a5c66de0/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.NamespaceFuture.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.NamespaceFuture.html
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.NamespaceFuture.html
index c10cfbf..a3e2f4a 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.NamespaceFuture.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.NamespaceFuture.html
@@ -3371,7 +3371,7 @@
3363private V result = null;
3364
3365private final HBaseAdmin admin;
-3366private final Long procId;
+3366protected final Long procId;
3367
3368public ProcedureFuture(final
HBaseAdmin admin, final Long procId) {
3369 this.admin = admin;
@@ -3653,653 +3653,651 @@
3645 * @return a description of the
operation
3646 */
3647protected String getDescription()
{
-3648 return "Operation: " +
getOperationType() + ", "
-3649 + "Table Name: " +
tableName.getNameWithNamespaceInclAsString();
-3650
-3651}
-3652
-3653protected abstract class
TableWaitForStateCallable implements WaitForStateCallable {
-3654 @Override
-3655 public void
throwInterruptedException() throws InterruptedIOException {
-3656throw new
InterruptedIOException("Interrupted while waiting for operation: "
-3657+ getOperationType() + " on
table: " + tableName.getNameWithNamespaceInclAsString());
-3658 }
-3659
-3660 @Override
-3661 public void
throwTimeoutException(long elapsedTime) throws TimeoutException {
-3662throw new TimeoutException("The
operation: " + getOperationType() + " on table: " +
-3663tableName.getNameAsString()
+ " has not completed after " + elapsedTime + "ms");
-3664 }
-3665}
-3666
-3667@Override
-3668protected V
postOperationResult(final V result, final long deadlineTs)
-3669throws IOException,
TimeoutException {
-3670 LOG.info(getDescription() + "
completed");
-3671 return
super.postOperationResult(result, deadlineTs);
-3672}
-3673
-3674@Override
-3675protected V
postOperationFailure(final IOException exception, final long deadlineTs)
-3676throws IOException,
TimeoutException {
-3677 LOG.info(getDescription() + "
failed with " + exception.getMessage());
-3678 return
super.postOperationFailure(exception, deadlineTs);
-3679}
-3680
-3681protected void
waitForTableEnabled(final long deadlineTs)
-3682throws IOException,
TimeoutException {
-3683 waitForState(deadlineTs, new
TableWaitForStateCallable() {
-3684@Override
-3685public boolean checkState(int
tries) throws IOException {
-3686 try {
-3687if
(getAdmin().isTableAvailable(tableName)) {
-3688 return true;
-3689}
-3690 } catch
(TableNotFoundException tnfe) {
-3691LOG.debug("Table " +
tableName.getNameWithNamespaceInclAsString()
-3692+ " was not enabled,
sleeping. tries=" + tries);
-3693 }
-3694 return false;
-3695}
-3696 });
-3697}
-3698
-3699protected void
waitForTableDisabled(final long deadlineTs)
-3700throws IOException,
TimeoutException {
-3701 waitForState(deadlineTs, new
TableWaitForStateCallable() {
-3702@Override
-3703public boolean checkState(int
tries) throws IOException {
-3704 return
getAdmin().isTableDisabled(tableName);
-3705}
-3706 });
-3707}
-3708
-3709protected void
waitTableNotFound(final long deadlineTs)
-3710throws IOException,
TimeoutException {
-3711 waitForState(deadlineTs, new
TableWaitForStateCallable() {
-3712@Override
-3713public boolean checkState(int
tries) throws IOException {
-3714 return
!getAdmin().tableExists(tableName);
-3715}
-3716 });
-3717}
-3718
-3719protected void
waitForSchemaUpdate(final long deadlineTs)
-3720throws IOException,
TimeoutException {
-3721 waitForState(deadlineTs, new
TableWaitForStateCallable() {
-3722@Override
-3723public boolean checkState(int
tries) throws IOException {
-3724 return
getAdmin().getAlterStatus(tableName).getFirst() == 0;
-3725}
-3726 });
-3727}
-3728
-3729protected void
waitForAllRegionsOnline(final long deadlineTs, final byte[][] splitKeys)
-3730throws IOException,
TimeoutException {
-3731 final TableDescriptor desc =
getTableDescriptor();
-3732 final AtomicInteger actualRegCount
= new AtomicInteger(0);
-3733 final MetaTableAccessor.Visitor
visitor = new MetaTableAccessor.Visitor() {
-3734@Override
-3735public boolean visit(Result
rowResult) throws IOException {
-3736
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb5d2c62/devapidocs/org/apache/hadoop/hbase/master/procedure/PeerProcedureInterface.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/procedure/PeerProcedureInterface.html
b/devapidocs/org/apache/hadoop/hbase/master/procedure/PeerProcedureInterface.html
index cc07d09..5db3a2e 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/procedure/PeerProcedureInterface.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/procedure/PeerProcedureInterface.html
@@ -101,7 +101,7 @@ var activeTableTab = "activeTableTab";
All Known Implementing Classes:
-AbstractPeerProcedure, AddPeerProcedure, DisablePeerProcedure, EnablePeerProcedure, ModifyPeerProcedure, RefreshPeerProcedure, RemovePeerProcedure, UpdatePeerConfigProcedure
+AbstractPeerProcedure, AddPeerProcedure, DisablePeerProcedure, EnablePeerProcedure, ModifyPeerProcedure, RecoverStandbyProcedure, RefreshPeerProcedure, RemovePeerProcedure, SyncReplicationReplayWALProcedure,
SyncReplicationReplayWALRemoteProcedure,
TransitPeerSyncReplicationStateProcedure,
UpdatePeerConfigProcedure
@@ -173,7 +173,7 @@ public interface
getPeerId
-https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">StringgetPeerId()
+https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">StringgetPeerId()
@@ -182,7 +182,7 @@ public interface
getPeerOperationType
-PeerProcedureInterface.PeerOperationTypegetPeerOperationType()
+PeerProcedureInterface.PeerOperationTypegetPeerOperationType()
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb5d2c62/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
index 792aeaf..f98ae5c 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/MasterProcedureEnv.html
@@ -2093,57 +2093,93 @@
RefreshPeerProcedure.abort(MasterProcedureEnvenv)
+protected boolean
+SyncReplicationReplayWALRemoteProcedure.abort(MasterProcedureEnvenv)
+
+
protected Procedure.LockState
AbstractPeerProcedure.acquireLock(MasterProcedureEnvenv)
-
+
private void
RefreshPeerProcedure.complete(MasterProcedureEnvenv,
https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
title="class or interface in
java.lang">Throwableerror)
+
+private void
+SyncReplicationReplayWALRemoteProcedure.complete(MasterProcedureEnvenv,
+https://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
title="class or interface in
java.lang">Throwableerror)
+
+private void
+TransitPeerSyncReplicationStateProcedure.createDirForRemoteWAL(MasterProcedureEnvenv)
+
+
protected ProcedureMasterProcedureEnv[]
RefreshPeerProcedure.execute(MasterProcedureEnvenv)
+
+protected ProcedureMasterProcedureEnv[]
+SyncReplicationReplayWALRemoteProcedure.execute(MasterProcedureEnvenv)
+
protected StateMachineProcedure.Flow
ModifyPeerProcedure.executeFromState(MasterProcedureEnvenv,
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.PeerModificationStatestate)
+protected StateMachineProcedure.Flow
+TransitPeerSyncReplicationStateProcedure.executeFromState(MasterProcedureEnvenv,
+
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.PeerSyncReplicationStateTransitionStatestate)
+
+
+protected StateMachineProcedure.Flow
+RecoverStandbyProcedure.executeFromState(MasterProcedureEnvenv,
+
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.RecoverStandbyStatestate)
+
+
+protected StateMachineProcedure.Flow
+SyncReplicationReplayWALProcedure.executeFromState(MasterProcedureEnvenv,
+
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.SyncReplicationReplayWALStatestate)
+
+
protected boolean
AbstractPeerProcedure.hasLock(MasterProcedureEnvenv)
-
+
protected boolean
AbstractPeerProcedure.holdLock(MasterProcedureEnvenv)
-
+
protected void
AddPeerProcedure.postPeerModification(MasterProcedureEnvenv)
-
+
protected void
EnablePeerProcedure.postPeerModification(MasterProcedureEnvenv)
-
+
protected void
UpdatePeerConfigProcedure.postPeerModification(MasterProcedureEnvenv)
-
+
protected abstract void
ModifyPeerProcedure.postPeerModification(MasterProcedureEnvenv)
Called before we finish the procedure.
-
+
protected void
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/55ce8d97/devapidocs/org/apache/hadoop/hbase/regionserver/Region.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/Region.html
b/devapidocs/org/apache/hadoop/hbase/regionserver/Region.html
index de1962a..db0fab9 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/Region.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/Region.html
@@ -18,7 +18,7 @@
catch(err) {
}
//-->
-var methods =
{"i0":6,"i1":6,"i2":18,"i3":6,"i4":18,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6,"i20":6,"i21":6,"i22":6,"i23":6,"i24":6,"i25":6,"i26":6,"i27":6,"i28":6,"i29":6,"i30":6,"i31":6,"i32":6,"i33":6,"i34":6,"i35":6,"i36":6,"i37":6,"i38":6,"i39":6,"i40":6,"i41":6,"i42":6,"i43":6,"i44":6,"i45":38,"i46":38,"i47":38,"i48":6,"i49":6,"i50":6,"i51":6,"i52":6,"i53":6,"i54":6,"i55":6};
+var methods =
{"i0":6,"i1":6,"i2":18,"i3":6,"i4":18,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6,"i20":6,"i21":6,"i22":6,"i23":6,"i24":6,"i25":6,"i26":6,"i27":6,"i28":6,"i29":6,"i30":6,"i31":6,"i32":6,"i33":6,"i34":6,"i35":6,"i36":6,"i37":6,"i38":6,"i39":6,"i40":6,"i41":6,"i42":6,"i43":6,"i44":6,"i45":6,"i46":38,"i47":38,"i48":38,"i49":6,"i50":6,"i51":6,"i52":6,"i53":6,"i54":6,"i55":6,"i56":6};
var tabs = {65535:["t0","All Methods"],2:["t2","Instance
Methods"],4:["t3","Abstract Methods"],16:["t5","Default
Methods"],32:["t6","Deprecated Methods"]};
var altColor = "altColor";
var rowColor = "rowColor";
@@ -278,69 +278,73 @@ extends
long
-getDataInMemoryWithoutWAL()
+getCpRequestsCount()
long
-getEarliestFlushTimeForAllStores()
+getDataInMemoryWithoutWAL()
long
-getFilteredReadRequestsCount()
+getEarliestFlushTimeForAllStores()
long
-getMaxFlushedSeqId()
+getFilteredReadRequestsCount()
+long
+getMaxFlushedSeqId()
+
+
https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
title="class or interface in java.util">Mapbyte[],https://docs.oracle.com/javase/8/docs/api/java/lang/Long.html?is-external=true;
title="class or interface in java.lang">Long
getMaxStoreSeqId()
-
+
long
getMemStoreDataSize()
-
+
long
getMemStoreHeapSize()
-
+
long
getMemStoreOffHeapSize()
-
+
long
getNumMutationsWithoutWAL()
-
+
long
getOldestHfileTs(booleanmajorCompactionOnly)
This can be used to determine the last time all files of
this region were major compacted.
-
+
long
getReadRequestsCount()
-
+
RegionInfo
getRegionInfo()
-
+
Region.RowLock
getRowLock(byte[]row,
booleanreadLock)
Get a row lock for the specified row.
-
+
RegionScanner
getScanner(Scanscan)
Return an iterator that scans over the HRegion, returning
the indicated
columns and rows specified by the Scan.
-
+
RegionScanner
getScanner(Scanscan,
https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListKeyValueScanneradditionalScanners)
@@ -348,67 +352,67 @@ extends Scan.
-
+
Store
getStore(byte[]family)
Return the Store for the given family
-
+
https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">Listhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
getStoreFileList(byte[][]columns)
-
+
https://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">List? extends Store
getStores()
Return the list of Stores managed by this region
-
+
TableDescriptor
getTableDescriptor()
-
+
long
getWriteRequestsCount()
-
+
Result
increment(Incrementincrement)
Perform one or more increment operations on a row.
-
+
boolean
isAvailable()
-
+
boolean
isClosed()
-
+
boolean
isClosing()
-
+
boolean
isMergeable()
-
+
boolean
isReadOnly()
-
+
boolean
isSplittable()
-
+
void
mutateRow(RowMutationsmutations)
Performs multiple mutations atomically on a single
row.
-
+
void
mutateRowsWithLocks(https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
title="class or interface in java.util">CollectionMutationmutations,
https://docs.oracle.com/javase/8/docs/api/java/util/Collection.html?is-external=true;
title="class or interface in
java.util">Collectionbyte[]rowsToLock,
@@ -417,7 +421,7 @@ extends Perform atomic mutations within the region.
-
+
void
processRowsWithLocks(RowProcessor?,?processor)
Deprecated.
@@ -426,7 +430,7 @@ extends
+
void
processRowsWithLocks(RowProcessor?,?processor,
longnonceGroup,
@@ -437,7 +441,7 @@ extends
+
void
processRowsWithLocks(RowProcessor?,?processor,
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/2b11656f/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html
new file mode 100644
index 000..546d2b6
--- /dev/null
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.TableDescriptorGetter.html
@@ -0,0 +1,3793 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+Source code
+
+
+
+
+001/**
+002 * Licensed to the Apache Software
Foundation (ASF) under one
+003 * or more contributor license
agreements. See the NOTICE file
+004 * distributed with this work for
additional information
+005 * regarding copyright ownership. The
ASF licenses this file
+006 * to you under the Apache License,
Version 2.0 (the
+007 * "License"); you may not use this file
except in compliance
+008 * with the License. You may obtain a
copy of the License at
+009 *
+010 *
http://www.apache.org/licenses/LICENSE-2.0
+011 *
+012 * Unless required by applicable law or
agreed to in writing, software
+013 * distributed under the License is
distributed on an "AS IS" BASIS,
+014 * WITHOUT WARRANTIES OR CONDITIONS OF
ANY KIND, either express or implied.
+015 * See the License for the specific
language governing permissions and
+016 * limitations under the License.
+017 */
+018package org.apache.hadoop.hbase.master;
+019
+020import static
org.apache.hadoop.hbase.HConstants.HBASE_MASTER_LOGCLEANER_PLUGINS;
+021
+022import com.google.protobuf.Descriptors;
+023import com.google.protobuf.Service;
+024import java.io.IOException;
+025import java.io.InterruptedIOException;
+026import java.lang.reflect.Constructor;
+027import
java.lang.reflect.InvocationTargetException;
+028import java.net.InetAddress;
+029import java.net.InetSocketAddress;
+030import java.net.UnknownHostException;
+031import java.util.ArrayList;
+032import java.util.Arrays;
+033import java.util.Collection;
+034import java.util.Collections;
+035import java.util.Comparator;
+036import java.util.EnumSet;
+037import java.util.HashMap;
+038import java.util.Iterator;
+039import java.util.List;
+040import java.util.Map;
+041import java.util.Map.Entry;
+042import java.util.Objects;
+043import java.util.Optional;
+044import java.util.Set;
+045import
java.util.concurrent.ExecutionException;
+046import java.util.concurrent.Future;
+047import java.util.concurrent.TimeUnit;
+048import
java.util.concurrent.TimeoutException;
+049import
java.util.concurrent.atomic.AtomicInteger;
+050import
java.util.concurrent.atomic.AtomicReference;
+051import java.util.function.Function;
+052import java.util.regex.Pattern;
+053import java.util.stream.Collectors;
+054import javax.servlet.ServletException;
+055import javax.servlet.http.HttpServlet;
+056import
javax.servlet.http.HttpServletRequest;
+057import
javax.servlet.http.HttpServletResponse;
+058import
org.apache.commons.lang3.StringUtils;
+059import
org.apache.hadoop.conf.Configuration;
+060import org.apache.hadoop.fs.Path;
+061import
org.apache.hadoop.hbase.ClusterId;
+062import
org.apache.hadoop.hbase.ClusterMetrics;
+063import
org.apache.hadoop.hbase.ClusterMetrics.Option;
+064import
org.apache.hadoop.hbase.ClusterMetricsBuilder;
+065import
org.apache.hadoop.hbase.DoNotRetryIOException;
+066import
org.apache.hadoop.hbase.HBaseIOException;
+067import
org.apache.hadoop.hbase.HBaseInterfaceAudience;
+068import
org.apache.hadoop.hbase.HConstants;
+069import
org.apache.hadoop.hbase.InvalidFamilyOperationException;
+070import
org.apache.hadoop.hbase.MasterNotRunningException;
+071import
org.apache.hadoop.hbase.MetaTableAccessor;
+072import
org.apache.hadoop.hbase.NamespaceDescriptor;
+073import
org.apache.hadoop.hbase.PleaseHoldException;
+074import
org.apache.hadoop.hbase.ReplicationPeerNotFoundException;
+075import
org.apache.hadoop.hbase.ScheduledChore;
+076import
org.apache.hadoop.hbase.ServerName;
+077import
org.apache.hadoop.hbase.TableDescriptors;
+078import
org.apache.hadoop.hbase.TableName;
+079import
org.apache.hadoop.hbase.TableNotDisabledException;
+080import
org.apache.hadoop.hbase.TableNotFoundException;
+081import
org.apache.hadoop.hbase.UnknownRegionException;
+082import
org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
+083import
org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
+084import
org.apache.hadoop.hbase.client.MasterSwitchType;
+085import
org.apache.hadoop.hbase.client.RegionInfo;
+086import
org.apache.hadoop.hbase.client.RegionInfoBuilder;
+087import
org.apache.hadoop.hbase.client.Result;
+088import
org.apache.hadoop.hbase.client.TableDescriptor;
+089import
org.apache.hadoop.hbase.client.TableDescriptorBuilder;
+090import
org.apache.hadoop.hbase.client.TableState;
+091import
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/65565d77/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.FlushedSequenceIdFlusher.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.FlushedSequenceIdFlusher.html
b/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.FlushedSequenceIdFlusher.html
new file mode 100644
index 000..ba6e4aa
--- /dev/null
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.FlushedSequenceIdFlusher.html
@@ -0,0 +1,1244 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+Source code
+
+
+
+
+001/*
+002 *
+003 * Licensed to the Apache Software
Foundation (ASF) under one
+004 * or more contributor license
agreements. See the NOTICE file
+005 * distributed with this work for
additional information
+006 * regarding copyright ownership. The
ASF licenses this file
+007 * to you under the Apache License,
Version 2.0 (the
+008 * "License"); you may not use this file
except in compliance
+009 * with the License. You may obtain a
copy of the License at
+010 *
+011 *
http://www.apache.org/licenses/LICENSE-2.0
+012 *
+013 * Unless required by applicable law or
agreed to in writing, software
+014 * distributed under the License is
distributed on an "AS IS" BASIS,
+015 * WITHOUT WARRANTIES OR CONDITIONS OF
ANY KIND, either express or implied.
+016 * See the License for the specific
language governing permissions and
+017 * limitations under the License.
+018 */
+019package org.apache.hadoop.hbase.master;
+020
+021import static
org.apache.hadoop.hbase.util.CollectionUtils.computeIfAbsent;
+022
+023import java.io.IOException;
+024import java.net.InetAddress;
+025import java.util.ArrayList;
+026import java.util.Collections;
+027import java.util.HashMap;
+028import java.util.Iterator;
+029import java.util.List;
+030import java.util.Map;
+031import java.util.Map.Entry;
+032import java.util.Set;
+033import
java.util.concurrent.ConcurrentNavigableMap;
+034import
java.util.concurrent.ConcurrentSkipListMap;
+035import
java.util.concurrent.CopyOnWriteArrayList;
+036import
java.util.concurrent.atomic.AtomicBoolean;
+037import java.util.function.Predicate;
+038import
org.apache.hadoop.conf.Configuration;
+039import
org.apache.hadoop.fs.FSDataInputStream;
+040import
org.apache.hadoop.fs.FSDataOutputStream;
+041import org.apache.hadoop.fs.FileSystem;
+042import org.apache.hadoop.fs.Path;
+043import
org.apache.hadoop.hbase.ClockOutOfSyncException;
+044import
org.apache.hadoop.hbase.HConstants;
+045import
org.apache.hadoop.hbase.NotServingRegionException;
+046import
org.apache.hadoop.hbase.RegionMetrics;
+047import
org.apache.hadoop.hbase.ScheduledChore;
+048import
org.apache.hadoop.hbase.ServerMetrics;
+049import
org.apache.hadoop.hbase.ServerMetricsBuilder;
+050import
org.apache.hadoop.hbase.ServerName;
+051import
org.apache.hadoop.hbase.YouAreDeadException;
+052import
org.apache.hadoop.hbase.client.ClusterConnection;
+053import
org.apache.hadoop.hbase.client.RegionInfo;
+054import
org.apache.hadoop.hbase.client.RetriesExhaustedException;
+055import
org.apache.hadoop.hbase.ipc.HBaseRpcController;
+056import
org.apache.hadoop.hbase.ipc.RpcControllerFactory;
+057import
org.apache.hadoop.hbase.master.assignment.RegionStates;
+058import
org.apache.hadoop.hbase.monitoring.MonitoredTask;
+059import
org.apache.hadoop.hbase.regionserver.HRegionServer;
+060import
org.apache.hadoop.hbase.util.Bytes;
+061import
org.apache.hadoop.hbase.util.FSUtils;
+062import
org.apache.hadoop.hbase.zookeeper.ZKUtil;
+063import
org.apache.hadoop.hbase.zookeeper.ZKWatcher;
+064import
org.apache.yetus.audience.InterfaceAudience;
+065import
org.apache.zookeeper.KeeperException;
+066import org.slf4j.Logger;
+067import org.slf4j.LoggerFactory;
+068
+069import
org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
+070import
org.apache.hbase.thirdparty.com.google.protobuf.ByteString;
+071import
org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations;
+072
+073import
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
+074import
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService;
+075import
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionStoreSequenceIds;
+076import
org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.StoreSequenceId;
+077import
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.FlushedRegionSequenceId;
+078import
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.FlushedSequenceId;
+079import
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.FlushedStoreSequenceId;
+080import
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionServerStartupRequest;
+081
+082/**
+083 * The ServerManager class manages info
about region servers.
+084 * p
+085
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e11cf2cb/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/WALCellCodec.StatelessUncompressor.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/WALCellCodec.StatelessUncompressor.html
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/WALCellCodec.StatelessUncompressor.html
new file mode 100644
index 000..9df0225
--- /dev/null
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/WALCellCodec.StatelessUncompressor.html
@@ -0,0 +1,476 @@
+http://www.w3.org/TR/html4/loose.dtd;>
+
+
+Source code
+
+
+
+
+001/*
+002 * Licensed to the Apache Software
Foundation (ASF) under one
+003 * or more contributor license
agreements. See the NOTICE file
+004 * distributed with this work for
additional information
+005 * regarding copyright ownership. The
ASF licenses this file
+006 * to you under the Apache License,
Version 2.0 (the
+007 * "License"); you may not use this file
except in compliance
+008 * with the License. You may obtain a
copy of the License at
+009 *
+010 *
http://www.apache.org/licenses/LICENSE-2.0
+011 *
+012 * Unless required by applicable law or
agreed to in writing, software
+013 * distributed under the License is
distributed on an "AS IS" BASIS,
+014 * WITHOUT WARRANTIES OR CONDITIONS OF
ANY KIND, either express or implied.
+015 * See the License for the specific
language governing permissions and
+016 * limitations under the License.
+017 */
+018package
org.apache.hadoop.hbase.regionserver.wal;
+019
+020import java.io.ByteArrayOutputStream;
+021import java.io.IOException;
+022import java.io.InputStream;
+023import java.io.OutputStream;
+024
+025import
org.apache.hadoop.conf.Configuration;
+026import org.apache.hadoop.hbase.Cell;
+027import
org.apache.hadoop.hbase.CellUtil;
+028import
org.apache.hadoop.hbase.HBaseInterfaceAudience;
+029import
org.apache.hadoop.hbase.PrivateCellUtil;
+030import
org.apache.hadoop.hbase.KeyValue;
+031import
org.apache.hadoop.hbase.KeyValueUtil;
+032import
org.apache.yetus.audience.InterfaceAudience;
+033import
org.apache.hadoop.hbase.codec.BaseDecoder;
+034import
org.apache.hadoop.hbase.codec.BaseEncoder;
+035import
org.apache.hadoop.hbase.codec.Codec;
+036import
org.apache.hadoop.hbase.codec.KeyValueCodecWithTags;
+037import
org.apache.hadoop.hbase.io.ByteBuffInputStream;
+038import
org.apache.hadoop.hbase.io.ByteBufferWriter;
+039import
org.apache.hadoop.hbase.io.ByteBufferWriterOutputStream;
+040import
org.apache.hadoop.hbase.io.util.Dictionary;
+041import
org.apache.hadoop.hbase.io.util.StreamUtils;
+042import
org.apache.hadoop.hbase.nio.ByteBuff;
+043import
org.apache.hadoop.hbase.util.ByteBufferUtils;
+044import
org.apache.hadoop.hbase.util.Bytes;
+045import
org.apache.hadoop.hbase.util.ReflectionUtils;
+046import org.apache.hadoop.io.IOUtils;
+047
+048import
org.apache.hbase.thirdparty.com.google.protobuf.ByteString;
+049import
org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations;
+050
+051
+052/**
+053 * Compression in this class is lifted
off Compressor/KeyValueCompression.
+054 * This is a pure coincidence... they are
independent and don't have to be compatible.
+055 *
+056 * This codec is used at server side for
writing cells to WAL as well as for sending edits
+057 * as part of the distributed splitting
process.
+058 */
+059@InterfaceAudience.LimitedPrivate({HBaseInterfaceAudience.COPROC,
+060 HBaseInterfaceAudience.PHOENIX,
HBaseInterfaceAudience.CONFIG})
+061public class WALCellCodec implements
Codec {
+062 /** Configuration key for the class to
use when encoding cells in the WAL */
+063 public static final String
WAL_CELL_CODEC_CLASS_KEY = "hbase.regionserver.wal.codec";
+064
+065 protected final CompressionContext
compression;
+066
+067 /**
+068 * bAll subclasses must
implement a no argument constructor/b
+069 */
+070 public WALCellCodec() {
+071this.compression = null;
+072 }
+073
+074 /**
+075 * Default constructor - ball
subclasses must implement a constructor with this signature /b
+076 * if they are to be dynamically loaded
from the {@link Configuration}.
+077 * @param conf configuration to
configure ttthis/tt
+078 * @param compression compression the
codec should support, can be ttnull/tt to indicate no
+079 * compression
+080 */
+081 public WALCellCodec(Configuration conf,
CompressionContext compression) {
+082this.compression = compression;
+083 }
+084
+085 public static String
getWALCellCodecClass(Configuration conf) {
+086return
conf.get(WAL_CELL_CODEC_CLASS_KEY, WALCellCodec.class.getName());
+087 }
+088
+089 /**
+090 * Create and setup a {@link
WALCellCodec} from the {@code cellCodecClsName} and
+091 * CompressionContext, if {@code
cellCodecClsName} is specified.
+092 * Otherwise Cell Codec classname is
read from
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/3469cbc0/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.DelayedTask.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.DelayedTask.html
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.DelayedTask.html
index 594ef24..17d5c40 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.DelayedTask.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.DelayedTask.html
@@ -170,241 +170,242 @@
162 }
163
164 /**
-165 * Add a remote rpc. Be sure to check
result for successful add.
+165 * Add a remote rpc.
166 * @param key the node identifier
-167 * @return True if we successfully
added the operation.
-168 */
-169 public boolean addOperationToNode(final
TRemote key, RemoteProcedure rp) {
+167 */
+168 public void addOperationToNode(final
TRemote key, RemoteProcedure rp)
+169 throws
NullTargetServerDispatchException, NoServerDispatchException,
NoNodeDispatchException {
170if (key == null) {
-171 // Key is remote server name. Be
careful. It could have been nulled by a concurrent
-172 // ServerCrashProcedure shutting
down outstanding RPC requests. See remoteCallFailed.
-173 return false;
-174}
-175assert key != null : "found null key
for node";
-176BufferNode node = nodeMap.get(key);
-177if (node == null) {
-178 return false;
-179}
-180node.add(rp);
-181// Check our node still in the map;
could have been removed by #removeNode.
-182return nodeMap.containsValue(node);
-183 }
-184
-185 /**
-186 * Remove a remote node
-187 * @param key the node identifier
-188 */
-189 public boolean removeNode(final TRemote
key) {
-190final BufferNode node =
nodeMap.remove(key);
-191if (node == null) return false;
-192node.abortOperationsInQueue();
-193return true;
-194 }
-195
-196 //
-197 // Task Helpers
-198 //
-199 protected FutureVoid
submitTask(CallableVoid task) {
-200return threadPool.submit(task);
-201 }
-202
-203 protected FutureVoid
submitTask(CallableVoid task, long delay, TimeUnit unit) {
-204final FutureTaskVoid
futureTask = new FutureTask(task);
-205timeoutExecutor.add(new
DelayedTask(futureTask, delay, unit));
-206return futureTask;
-207 }
-208
-209 protected abstract void
remoteDispatch(TRemote key, SetRemoteProcedure operations);
-210 protected abstract void
abortPendingOperations(TRemote key, SetRemoteProcedure operations);
-211
-212 /**
-213 * Data structure with reference to
remote operation.
-214 */
-215 public static abstract class
RemoteOperation {
-216private final RemoteProcedure
remoteProcedure;
-217
-218protected RemoteOperation(final
RemoteProcedure remoteProcedure) {
-219 this.remoteProcedure =
remoteProcedure;
-220}
-221
-222public RemoteProcedure
getRemoteProcedure() {
-223 return remoteProcedure;
-224}
-225 }
-226
-227 /**
-228 * Remote procedure reference.
-229 */
-230 public interface
RemoteProcedureTEnv, TRemote {
-231/**
-232 * For building the remote
operation.
-233 */
-234RemoteOperation remoteCallBuild(TEnv
env, TRemote remote);
-235
-236/**
-237 * Called when the executeProcedure
call is failed.
-238 */
-239void remoteCallFailed(TEnv env,
TRemote remote, IOException exception);
-240
-241/**
-242 * Called when RS tells the remote
procedure is succeeded through the
-243 * {@code reportProcedureDone}
method.
-244 */
-245void remoteOperationCompleted(TEnv
env);
-246
-247/**
-248 * Called when RS tells the remote
procedure is failed through the {@code reportProcedureDone}
-249 * method.
-250 */
-251void remoteOperationFailed(TEnv env,
RemoteProcedureException error);
-252 }
-253
-254 /**
-255 * Account of what procedures are
running on remote node.
-256 * @param TEnv
-257 * @param TRemote
-258 */
-259 public interface RemoteNodeTEnv,
TRemote {
-260TRemote getKey();
-261void add(RemoteProcedureTEnv,
TRemote operation);
-262void dispatch();
-263 }
-264
-265 protected
ArrayListMultimapClass?, RemoteOperation
buildAndGroupRequestByType(final TEnv env,
-266 final TRemote remote, final
SetRemoteProcedure remoteProcedures) {
-267final
ArrayListMultimapClass?, RemoteOperation requestByType =
ArrayListMultimap.create();
-268for (RemoteProcedure proc:
remoteProcedures) {
-269 RemoteOperation operation =
proc.remoteCallBuild(env, remote);
-270
requestByType.put(operation.getClass(),
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4df09ed9/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
--
diff --git
a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
index d2f8d9e..43fd3a1 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
@@ -947,674 +947,678 @@
static HBaseClassTestRule
-TestImmutableHTableDescriptor.CLASS_RULE
+TestTableDescriptorUtils.CLASS_RULE
static HBaseClassTestRule
-TestTableDescriptorBuilder.CLASS_RULE
+TestImmutableHTableDescriptor.CLASS_RULE
static HBaseClassTestRule
-TestBufferedMutatorParams.CLASS_RULE
+TestTableDescriptorBuilder.CLASS_RULE
static HBaseClassTestRule
-TestAsyncProcess.CLASS_RULE
+TestBufferedMutatorParams.CLASS_RULE
static HBaseClassTestRule
-TestClientScanner.CLASS_RULE
+TestAsyncProcess.CLASS_RULE
static HBaseClassTestRule
-TestClientNoCluster.CLASS_RULE
+TestClientScanner.CLASS_RULE
static HBaseClassTestRule
-TestAttributes.CLASS_RULE
+TestClientNoCluster.CLASS_RULE
static HBaseClassTestRule
-TestMetricsConnection.CLASS_RULE
+TestAttributes.CLASS_RULE
static HBaseClassTestRule
-TestSnapshotFromAdmin.CLASS_RULE
+TestMetricsConnection.CLASS_RULE
static HBaseClassTestRule
-TestDeleteTimeStamp.CLASS_RULE
+TestSnapshotFromAdmin.CLASS_RULE
static HBaseClassTestRule
-TestAsyncProcessWithRegionException.CLASS_RULE
+TestDeleteTimeStamp.CLASS_RULE
static HBaseClassTestRule
-TestHTableMultiplexerViaMocks.CLASS_RULE
+TestAsyncProcessWithRegionException.CLASS_RULE
static HBaseClassTestRule
-TestImmutableHRegionInfo.CLASS_RULE
+TestHTableMultiplexerViaMocks.CLASS_RULE
static HBaseClassTestRule
-TestDelayingRunner.CLASS_RULE
+TestImmutableHRegionInfo.CLASS_RULE
static HBaseClassTestRule
-TestReversedScannerCallable.CLASS_RULE
+TestDelayingRunner.CLASS_RULE
static HBaseClassTestRule
-TestCoprocessorDescriptor.CLASS_RULE
+TestReversedScannerCallable.CLASS_RULE
static HBaseClassTestRule
-TestClientExponentialBackoff.CLASS_RULE
+TestCoprocessorDescriptor.CLASS_RULE
static HBaseClassTestRule
-TestInterfaceAlign.CLASS_RULE
+TestClientExponentialBackoff.CLASS_RULE
static HBaseClassTestRule
-TestIncrement.CLASS_RULE
+TestInterfaceAlign.CLASS_RULE
static HBaseClassTestRule
-TestColumnFamilyDescriptorBuilder.CLASS_RULE
+TestIncrement.CLASS_RULE
static HBaseClassTestRule
-TestOperation.CLASS_RULE
+TestColumnFamilyDescriptorBuilder.CLASS_RULE
static HBaseClassTestRule
-TestMutation.CLASS_RULE
+TestOperation.CLASS_RULE
static HBaseClassTestRule
-TestScan.CLASS_RULE
+TestMutation.CLASS_RULE
static HBaseClassTestRule
-TestRetriesExhaustedWithDetailsException.CLASS_RULE
+TestScan.CLASS_RULE
static HBaseClassTestRule
-TestSimpleRequestController.CLASS_RULE
+TestRetriesExhaustedWithDetailsException.CLASS_RULE
static HBaseClassTestRule
-TestRowComparator.CLASS_RULE
+TestSimpleRequestController.CLASS_RULE
static HBaseClassTestRule
-TestProcedureFuture.CLASS_RULE
+TestRowComparator.CLASS_RULE
static HBaseClassTestRule
-TestRegionInfoDisplay.CLASS_RULE
+TestProcedureFuture.CLASS_RULE
static HBaseClassTestRule
-TestPutDotHas.CLASS_RULE
+TestRegionInfoDisplay.CLASS_RULE
static HBaseClassTestRule
-TestGet.CLASS_RULE
+TestPutDotHas.CLASS_RULE
static HBaseClassTestRule
-TestImmutableHColumnDescriptor.CLASS_RULE
+TestGet.CLASS_RULE
static HBaseClassTestRule
-TestBufferedMutator.CLASS_RULE
+TestImmutableHColumnDescriptor.CLASS_RULE
static HBaseClassTestRule
-TestAsyncNamespaceAdminApi.CLASS_RULE
+TestBufferedMutator.CLASS_RULE
static HBaseClassTestRule
-TestRawAsyncScanCursor.CLASS_RULE
+TestAsyncNamespaceAdminApi.CLASS_RULE
static HBaseClassTestRule
-TestReplicasClient.CLASS_RULE
+TestRawAsyncScanCursor.CLASS_RULE
static HBaseClassTestRule
-TestGetProcedureResult.CLASS_RULE
+TestReplicasClient.CLASS_RULE
static HBaseClassTestRule
-TestAsyncReplicationAdminApi.CLASS_RULE
+TestGetProcedureResult.CLASS_RULE
static HBaseClassTestRule
-TestAsyncSnapshotAdminApi.CLASS_RULE
+TestAsyncReplicationAdminApi.CLASS_RULE
static HBaseClassTestRule
-TestClientTimeouts.CLASS_RULE
+TestAsyncSnapshotAdminApi.CLASS_RULE
static HBaseClassTestRule
-TestAsyncTable.CLASS_RULE
+TestClientTimeouts.CLASS_RULE
static HBaseClassTestRule
-TestCIPutRpcTimeout.CLASS_RULE
+TestAsyncTable.CLASS_RULE
static HBaseClassTestRule
-TestRestoreSnapshotFromClientWithRegionReplicas.CLASS_RULE
+TestCIPutRpcTimeout.CLASS_RULE
static HBaseClassTestRule
-TestLimitedScanWithFilter.CLASS_RULE
+TestRestoreSnapshotFromClientWithRegionReplicas.CLASS_RULE
static HBaseClassTestRule
-TestDropTimeoutRequest.CLASS_RULE
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/883dde2f/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html
b/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html
index 28539e5..c3f794b 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/wal/WALFactory.Providers.html
@@ -55,404 +55,406 @@
047 * implementations:
048 * ul
049 *
liemdefaultProvider/em : whatever provider is standard
for the hbase version. Currently
-050 *
"filesystem"/li
-051 *
liemfilesystem/em : a provider that will run on top of
an implementation of the Hadoop
-052 * FileSystem
interface, normally HDFS./li
-053 *
liemmultiwal/em : a provider that will use multiple
"filesystem" wal instances per region
-054 *
server./li
-055 * /ul
-056 *
-057 * Alternatively, you may provide a
custom implementation of {@link WALProvider} by class name.
-058 */
-059@InterfaceAudience.Private
-060public class WALFactory {
-061
-062 private static final Logger LOG =
LoggerFactory.getLogger(WALFactory.class);
+050 *
"asyncfs"/li
+051 *
liemasyncfs/em : a provider that will run on top of an
implementation of the Hadoop
+052 * FileSystem
interface via an asynchronous client./li
+053 *
liemfilesystem/em : a provider that will run on top of
an implementation of the Hadoop
+054 * FileSystem
interface via HDFS's synchronous DFSClient./li
+055 *
liemmultiwal/em : a provider that will use multiple
"filesystem" wal instances per region
+056 *
server./li
+057 * /ul
+058 *
+059 * Alternatively, you may provide a
custom implementation of {@link WALProvider} by class name.
+060 */
+061@InterfaceAudience.Private
+062public class WALFactory {
063
-064 /**
-065 * Maps between configuration names for
providers and implementation classes.
-066 */
-067 static enum Providers {
-068
defaultProvider(AsyncFSWALProvider.class),
-069filesystem(FSHLogProvider.class),
-070
multiwal(RegionGroupingProvider.class),
-071asyncfs(AsyncFSWALProvider.class);
-072
-073final Class? extends
WALProvider clazz;
-074Providers(Class? extends
WALProvider clazz) {
-075 this.clazz = clazz;
-076}
-077 }
-078
-079 public static final String WAL_PROVIDER
= "hbase.wal.provider";
-080 static final String
DEFAULT_WAL_PROVIDER = Providers.defaultProvider.name();
-081
-082 public static final String
META_WAL_PROVIDER = "hbase.wal.meta_provider";
-083 static final String
DEFAULT_META_WAL_PROVIDER = Providers.defaultProvider.name();
-084
-085 final String factoryId;
-086 private final WALProvider provider;
-087 // The meta updates are written to a
different wal. If this
-088 // regionserver holds meta regions,
then this ref will be non-null.
-089 // lazily intialized; most
RegionServers don't deal with META
-090 private final
AtomicReferenceWALProvider metaProvider = new
AtomicReference();
-091
-092 /**
-093 * Configuration-specified WAL Reader
used when a custom reader is requested
-094 */
-095 private final Class? extends
AbstractFSWALProvider.Reader logReaderClass;
-096
-097 /**
-098 * How long to attempt opening
in-recovery wals
-099 */
-100 private final int timeoutMillis;
-101
-102 private final Configuration conf;
+064 private static final Logger LOG =
LoggerFactory.getLogger(WALFactory.class);
+065
+066 /**
+067 * Maps between configuration names for
providers and implementation classes.
+068 */
+069 static enum Providers {
+070
defaultProvider(AsyncFSWALProvider.class),
+071filesystem(FSHLogProvider.class),
+072
multiwal(RegionGroupingProvider.class),
+073asyncfs(AsyncFSWALProvider.class);
+074
+075final Class? extends
WALProvider clazz;
+076Providers(Class? extends
WALProvider clazz) {
+077 this.clazz = clazz;
+078}
+079 }
+080
+081 public static final String WAL_PROVIDER
= "hbase.wal.provider";
+082 static final String
DEFAULT_WAL_PROVIDER = Providers.defaultProvider.name();
+083
+084 public static final String
META_WAL_PROVIDER = "hbase.wal.meta_provider";
+085 static final String
DEFAULT_META_WAL_PROVIDER = Providers.defaultProvider.name();
+086
+087 final String factoryId;
+088 private final WALProvider provider;
+089 // The meta updates are written to a
different wal. If this
+090 // regionserver holds meta regions,
then this ref will be non-null.
+091 // lazily intialized; most
RegionServers don't deal with META
+092 private final
AtomicReferenceWALProvider metaProvider = new
AtomicReference();
+093
+094 /**
+095 *
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7bcc960d/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTest.html
--
diff --git
a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTest.html
b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTest.html
index 3f8844b..cdb9398 100644
---
a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTest.html
+++
b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTest.html
@@ -140,2712 +140,2713 @@
132public class PerformanceEvaluation
extends Configured implements Tool {
133 static final String RANDOM_SEEK_SCAN =
"randomSeekScan";
134 static final String RANDOM_READ =
"randomRead";
-135 private static final Logger LOG =
LoggerFactory.getLogger(PerformanceEvaluation.class.getName());
-136 private static final ObjectMapper
MAPPER = new ObjectMapper();
-137 static {
-138
MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true);
-139 }
-140
-141 public static final String TABLE_NAME =
"TestTable";
-142 public static final String
FAMILY_NAME_BASE = "info";
-143 public static final byte[] FAMILY_ZERO
= Bytes.toBytes("info0");
-144 public static final byte[] COLUMN_ZERO
= Bytes.toBytes("" + 0);
-145 public static final int
DEFAULT_VALUE_LENGTH = 1000;
-146 public static final int ROW_LENGTH =
26;
-147
-148 private static final int ONE_GB = 1024
* 1024 * 1000;
-149 private static final int
DEFAULT_ROWS_PER_GB = ONE_GB / DEFAULT_VALUE_LENGTH;
-150 // TODO : should we make this
configurable
-151 private static final int TAG_LENGTH =
256;
-152 private static final DecimalFormat FMT
= new DecimalFormat("0.##");
-153 private static final MathContext CXT =
MathContext.DECIMAL64;
-154 private static final BigDecimal
MS_PER_SEC = BigDecimal.valueOf(1000);
-155 private static final BigDecimal
BYTES_PER_MB = BigDecimal.valueOf(1024 * 1024);
-156 private static final TestOptions
DEFAULT_OPTS = new TestOptions();
-157
-158 private static MapString,
CmdDescriptor COMMANDS = new TreeMap();
-159 private static final Path PERF_EVAL_DIR
= new Path("performance_evaluation");
-160
-161 static {
-162
addCommandDescriptor(AsyncRandomReadTest.class, "asyncRandomRead",
-163"Run async random read test");
-164
addCommandDescriptor(AsyncRandomWriteTest.class, "asyncRandomWrite",
-165"Run async random write test");
-166
addCommandDescriptor(AsyncSequentialReadTest.class, "asyncSequentialRead",
-167"Run async sequential read
test");
-168
addCommandDescriptor(AsyncSequentialWriteTest.class, "asyncSequentialWrite",
-169"Run async sequential write
test");
-170
addCommandDescriptor(AsyncScanTest.class, "asyncScan",
-171"Run async scan test (read every
row)");
-172
addCommandDescriptor(RandomReadTest.class, RANDOM_READ,
-173 "Run random read test");
-174
addCommandDescriptor(RandomSeekScanTest.class, RANDOM_SEEK_SCAN,
-175 "Run random seek and scan 100
test");
-176
addCommandDescriptor(RandomScanWithRange10Test.class, "scanRange10",
-177 "Run random seek scan with both
start and stop row (max 10 rows)");
-178
addCommandDescriptor(RandomScanWithRange100Test.class, "scanRange100",
-179 "Run random seek scan with both
start and stop row (max 100 rows)");
-180
addCommandDescriptor(RandomScanWithRange1000Test.class, "scanRange1000",
-181 "Run random seek scan with both
start and stop row (max 1000 rows)");
-182
addCommandDescriptor(RandomScanWithRange1Test.class, "scanRange1",
-183 "Run random seek scan with both
start and stop row (max 1 rows)");
-184
addCommandDescriptor(RandomWriteTest.class, "randomWrite",
-185 "Run random write test");
-186
addCommandDescriptor(SequentialReadTest.class, "sequentialRead",
-187 "Run sequential read test");
-188
addCommandDescriptor(SequentialWriteTest.class, "sequentialWrite",
-189 "Run sequential write test");
-190addCommandDescriptor(ScanTest.class,
"scan",
-191 "Run scan test (read every
row)");
-192
addCommandDescriptor(FilteredScanTest.class, "filterScan",
-193 "Run scan test using a filter to
find a specific row based on it's value " +
-194 "(make sure to use --rows=20)");
-195
addCommandDescriptor(IncrementTest.class, "increment",
-196 "Increment on each row; clients
overlap on keyspace so some concurrent operations");
-197
addCommandDescriptor(AppendTest.class, "append",
-198 "Append on each row; clients
overlap on keyspace so some concurrent operations");
-199
addCommandDescriptor(CheckAndMutateTest.class, "checkAndMutate",
-200 "CheckAndMutate on each row;
clients overlap on keyspace so some concurrent operations");
-201
addCommandDescriptor(CheckAndPutTest.class, "checkAndPut",
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/92a26cfb/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MemStoreSize.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MemStoreSize.html
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MemStoreSize.html
index cdc0ca3..f434f88 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MemStoreSize.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/MemStoreSize.html
@@ -27,109 +27,99 @@
019
020import
org.apache.hadoop.hbase.HBaseInterfaceAudience;
021import
org.apache.yetus.audience.InterfaceAudience;
-022
-023/**
-024 * Reports the data size part and total
heap space occupied by the MemStore.
-025 * Read-only.
+022/**
+023 * Data structure of three longs.
+024 * Convenient package in which to carry
current state of three counters.
+025 * pImmutable!/p
026 * @see MemStoreSizing
027 */
028@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.COPROC)
029public class MemStoreSize {
-030 // MemStore size tracks 3 sizes:
-031 // (1) data size: the aggregated size
of all key-value not including meta data such as
-032 // index, time range etc.
-033 // (2) heap size: the aggregated size
of all data that is allocated on-heap including all
-034 // key-values that reside on-heap and
the metadata that resides on-heap
-035 // (3) off-heap size: the aggregated
size of all data that is allocated off-heap including all
-036 // key-values that reside off-heap and
the metadata that resides off-heap
-037 //
-038 // 3 examples to illustrate their
usage:
-039 // Consider a store with 100MB of
key-values allocated on-heap and 20MB of metadata allocated
-040 // on-heap. The counters are 100MB,
120MB, 0, respectively.
-041 // Consider a store with 100MB of
key-values allocated off-heap and 20MB of metadata
-042 // allocated on-heap (e.g, CAM index).
The counters are 100MB, 20MB, 100MB, respectively.
-043 // Consider a store with 100MB of
key-values from which 95MB are allocated off-heap and 5MB
-044 // are allocated on-heap (e.g., due to
upserts) and 20MB of metadata from which 15MB allocated
-045 // off-heap (e.g, CCM index) and 5MB
allocated on-heap (e.g, CSLM index in active).
-046 // The counters are 100MB, 10MB,
110MB, respectively.
-047
-048 /**
-049 *'dataSize' tracks the Cell's data
bytes size alone (Key bytes, value bytes). A cell's data can
-050 * be in on heap or off heap area
depending on the MSLAB and its configuration to be using on heap
-051 * or off heap LABs
-052 */
-053 protected volatile long dataSize;
-054
-055 /** 'heapSize' tracks all Cell's heap
size occupancy. This will include Cell POJO heap overhead.
-056 * When Cells in on heap area, this
will include the cells data size as well.
-057 */
-058 protected volatile long heapSize;
-059
-060 /** off-heap size: the aggregated size
of all data that is allocated off-heap including all
-061 * key-values that reside off-heap and
the metadata that resides off-heap
-062 */
-063 protected volatile long offHeapSize;
-064
-065 public MemStoreSize() {
-066this(0L, 0L, 0L);
-067 }
-068
-069 public MemStoreSize(long dataSize, long
heapSize, long offHeapSize) {
-070this.dataSize = dataSize;
-071this.heapSize = heapSize;
-072this.offHeapSize = offHeapSize;
-073 }
-074
-075 protected MemStoreSize(MemStoreSize
memStoreSize) {
-076this.dataSize =
memStoreSize.dataSize;
-077this.heapSize =
memStoreSize.heapSize;
-078this.offHeapSize =
memStoreSize.offHeapSize;
-079 }
-080 public boolean isEmpty() {
-081return this.dataSize == 0
this.heapSize == 0 this.offHeapSize == 0;
+030 /**
+031 *'dataSize' tracks the Cell's data
bytes size alone (Key bytes, value bytes). A cell's data can
+032 * be in on heap or off heap area
depending on the MSLAB and its configuration to be using on
+033 * heap or off heap LABs
+034 */
+035 private final long dataSize;
+036
+037 /**'getHeapSize' tracks all Cell's heap
size occupancy. This will include Cell POJO heap overhead.
+038 * When Cells in on heap area, this
will include the cells data size as well.
+039 */
+040 private final long heapSize;
+041
+042 /** off-heap size: the aggregated size
of all data that is allocated off-heap including all
+043 * key-values that reside off-heap and
the metadata that resides off-heap
+044 */
+045 private final long offHeapSize;
+046
+047 /**
+048 * Package private constructor.
+049 */
+050 MemStoreSize() {
+051this(0L, 0L, 0L);
+052 }
+053
+054 /**
+055 * Package private constructor.
+056 */
+057 MemStoreSize(long dataSize, long
heapSize, long offHeapSize) {
+058this.dataSize = dataSize;
+059this.heapSize = heapSize;
+060this.offHeapSize = offHeapSize;
+061 }
+062
+063 /**
+064 * Package private constructor.
+065 */
+066
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomWriteTest.html
--
diff --git
a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomWriteTest.html
b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomWriteTest.html
index 2510283..418c60c 100644
---
a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomWriteTest.html
+++
b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomWriteTest.html
@@ -77,77 +77,77 @@
069import
org.apache.hadoop.hbase.client.RowMutations;
070import
org.apache.hadoop.hbase.client.Scan;
071import
org.apache.hadoop.hbase.client.Table;
-072import
org.apache.hadoop.hbase.filter.BinaryComparator;
-073import
org.apache.hadoop.hbase.filter.Filter;
-074import
org.apache.hadoop.hbase.filter.FilterAllFilter;
-075import
org.apache.hadoop.hbase.filter.FilterList;
-076import
org.apache.hadoop.hbase.filter.PageFilter;
-077import
org.apache.hadoop.hbase.filter.SingleColumnValueFilter;
-078import
org.apache.hadoop.hbase.filter.WhileMatchFilter;
-079import
org.apache.hadoop.hbase.io.compress.Compression;
-080import
org.apache.hadoop.hbase.io.encoding.DataBlockEncoding;
-081import
org.apache.hadoop.hbase.io.hfile.RandomDistribution;
-082import
org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil;
-083import
org.apache.hadoop.hbase.regionserver.BloomType;
-084import
org.apache.hadoop.hbase.regionserver.CompactingMemStore;
-085import
org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration;
-086import
org.apache.hadoop.hbase.trace.SpanReceiverHost;
-087import
org.apache.hadoop.hbase.trace.TraceUtil;
-088import
org.apache.hadoop.hbase.util.ByteArrayHashKey;
-089import
org.apache.hadoop.hbase.util.Bytes;
-090import
org.apache.hadoop.hbase.util.Hash;
-091import
org.apache.hadoop.hbase.util.MurmurHash;
-092import
org.apache.hadoop.hbase.util.Pair;
-093import
org.apache.hadoop.hbase.util.YammerHistogramUtils;
-094import
org.apache.hadoop.io.LongWritable;
-095import org.apache.hadoop.io.Text;
-096import org.apache.hadoop.mapreduce.Job;
-097import
org.apache.hadoop.mapreduce.Mapper;
-098import
org.apache.hadoop.mapreduce.lib.input.NLineInputFormat;
-099import
org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
-100import
org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer;
-101import org.apache.hadoop.util.Tool;
-102import
org.apache.hadoop.util.ToolRunner;
-103import
org.apache.htrace.core.ProbabilitySampler;
-104import org.apache.htrace.core.Sampler;
-105import
org.apache.htrace.core.TraceScope;
-106import
org.apache.yetus.audience.InterfaceAudience;
-107import org.slf4j.Logger;
-108import org.slf4j.LoggerFactory;
-109import
org.apache.hbase.thirdparty.com.google.common.base.MoreObjects;
-110import
org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder;
-111
-112/**
-113 * Script used evaluating HBase
performance and scalability. Runs a HBase
-114 * client that steps through one of a set
of hardcoded tests or 'experiments'
-115 * (e.g. a random reads test, a random
writes test, etc.). Pass on the
-116 * command-line which test to run and how
many clients are participating in
-117 * this experiment. Run {@code
PerformanceEvaluation --help} to obtain usage.
-118 *
-119 * pThis class sets up and runs
the evaluation programs described in
-120 * Section 7, iPerformance
Evaluation/i, of the a
-121 *
href="http://labs.google.com/papers/bigtable.html"Bigtable/a;
-122 * paper, pages 8-10.
-123 *
-124 * pBy default, runs as a
mapreduce job where each mapper runs a single test
-125 * client. Can also run as a
non-mapreduce, multithreaded application by
-126 * specifying {@code --nomapred}. Each
client does about 1GB of data, unless
-127 * specified otherwise.
-128 */
-129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS)
-130public class PerformanceEvaluation
extends Configured implements Tool {
-131 static final String RANDOM_SEEK_SCAN =
"randomSeekScan";
-132 static final String RANDOM_READ =
"randomRead";
-133 private static final Logger LOG =
LoggerFactory.getLogger(PerformanceEvaluation.class.getName());
-134 private static final ObjectMapper
MAPPER = new ObjectMapper();
-135 static {
-136
MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true);
-137 }
-138
-139 public static final String TABLE_NAME =
"TestTable";
-140 public static final byte[] FAMILY_NAME
= Bytes.toBytes("info");
-141 public static final byte [] COLUMN_ZERO
= Bytes.toBytes("" + 0);
-142 public static final byte []
QUALIFIER_NAME = COLUMN_ZERO;
+072import
org.apache.hadoop.hbase.client.metrics.ScanMetrics;
+073import
org.apache.hadoop.hbase.filter.BinaryComparator;
+074import
org.apache.hadoop.hbase.filter.Filter;
+075import
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/de18d468/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-frame.html
--
diff --git
a/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-frame.html
b/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-frame.html
index 7303f51..c604218 100644
---
a/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-frame.html
+++
b/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-frame.html
@@ -32,9 +32,10 @@
TestReplicationThrottler
TestReplicator
TestReplicator.FailureInjectingReplicationEndpointForTest
-TestReplicator.FailureInjectingReplicationEndpointForTest.FailureInjectingBlockingInterface
TestReplicator.ReplicationEndpointForTest
TestSerialReplicationChecker
+TestSerialReplicationEndpoint
+TestSerialReplicationEndpoint.TestEndpoint
TestSourceFSConfigurationProvider
TestWALEntrySinkFilter
TestWALEntrySinkFilter.DevNullConnection
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/de18d468/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-summary.html
--
diff --git
a/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-summary.html
b/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-summary.html
index 9b27b42..fd9dcc8 100644
---
a/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-summary.html
+++
b/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-summary.html
@@ -173,40 +173,44 @@
-TestReplicator.FailureInjectingReplicationEndpointForTest.FailureInjectingBlockingInterface
+TestReplicator.ReplicationEndpointForTest
-
-TestReplicator.ReplicationEndpointForTest
+
+TestSerialReplicationChecker
-TestSerialReplicationChecker
+TestSerialReplicationEndpoint
-TestSourceFSConfigurationProvider
+TestSerialReplicationEndpoint.TestEndpoint
+TestSourceFSConfigurationProvider
+
+
+
TestWALEntrySinkFilter
Simple test of sink-side wal entry filter facility.
-
+
TestWALEntrySinkFilter.DevNullConnection
A DevNull Connection whose only purpose is checking what
edits made it through.
-
+
TestWALEntrySinkFilter.IfTimeIsGreaterThanBOUNDARYWALEntrySinkFilterImpl
Simple filter that will filter out any entry wholse
writeTime is <= 5.
-
+
TestWALEntryStream
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/de18d468/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html
--
diff --git
a/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html
b/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html
index f7c6455..b47d481 100644
---
a/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html
+++
b/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html
@@ -94,6 +94,7 @@
org.apache.hadoop.hbase.replication.regionserver.TestReplicator.FailureInjectingReplicationEndpointForTest
+org.apache.hadoop.hbase.replication.regionserver.TestSerialReplicationEndpoint.TestEndpoint
@@ -138,13 +139,8 @@
org.apache.hadoop.hbase.replication.regionserver.TestReplicationSourceManager.DummyServer
(implements org.apache.hadoop.hbase.Server)
org.apache.hadoop.hbase.replication.regionserver.TestReplicationThrottler
-org.apache.hadoop.hbase.replication.regionserver.TestReplicator.FailureInjectingReplicationEndpointForTest.FailureInjectingBlockingInterface
(implements
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService.BlockingInterface)
-org.apache.hadoop.hbase.replication.regionserver.TestReplicator.ReplicationEndpointForTest.ReplicatorForTest
-
-org.apache.hadoop.hbase.replication.regionserver.TestReplicator.FailureInjectingReplicationEndpointForTest.FailureInjectingReplicatorForTest
-
-
org.apache.hadoop.hbase.replication.regionserver.TestSerialReplicationChecker
+org.apache.hadoop.hbase.replication.regionserver.TestSerialReplicationEndpoint
org.apache.hadoop.hbase.replication.regionserver.TestSourceFSConfigurationProvider (implements
org.apache.hadoop.hbase.replication.regionserver.SourceFSConfigurationProvider)
org.apache.hadoop.hbase.replication.regionserver.TestWALEntrySinkFilter
org.apache.hadoop.hbase.replication.regionserver.TestWALEntrySinkFilter.DevNullConnection
(implements org.apache.hadoop.hbase.client.Connection)
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/de18d468/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-use.html
--
diff --git
a/testdevapidocs/org/apache/hadoop/hbase/replication/regionserver/package-use.html
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d220bc5e/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.TableInfo.HDFSIntegrityFixer.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.TableInfo.HDFSIntegrityFixer.html
b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.TableInfo.HDFSIntegrityFixer.html
index 8302e28..c370eb9 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.TableInfo.HDFSIntegrityFixer.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/util/HBaseFsck.TableInfo.HDFSIntegrityFixer.html
@@ -2113,3031 +2113,3033 @@
2105
errors.reportError(ERROR_CODE.ORPHAN_TABLE_STATE,
2106tableName + " unable to
delete dangling table state " + tableState);
2107 }
-2108} else {
-2109
errors.reportError(ERROR_CODE.ORPHAN_TABLE_STATE,
-2110 tableName + " has dangling
table state " + tableState);
-2111}
-2112 }
-2113}
-2114// check that all tables have
states
-2115for (TableName tableName :
tablesInfo.keySet()) {
-2116 if (isTableIncluded(tableName)
!tableStates.containsKey(tableName)) {
-2117if (fixMeta) {
-2118
MetaTableAccessor.updateTableState(connection, tableName,
TableState.State.ENABLED);
-2119 TableState newState =
MetaTableAccessor.getTableState(connection, tableName);
-2120 if (newState == null) {
-2121
errors.reportError(ERROR_CODE.NO_TABLE_STATE,
-2122"Unable to change state
for table " + tableName + " in meta ");
-2123 }
-2124} else {
-2125
errors.reportError(ERROR_CODE.NO_TABLE_STATE,
-2126 tableName + " has no state
in meta ");
-2127}
-2128 }
-2129}
-2130 }
-2131
-2132 private void preCheckPermission()
throws IOException, AccessDeniedException {
-2133if
(shouldIgnorePreCheckPermission()) {
-2134 return;
-2135}
-2136
-2137Path hbaseDir =
FSUtils.getRootDir(getConf());
-2138FileSystem fs =
hbaseDir.getFileSystem(getConf());
-2139UserProvider userProvider =
UserProvider.instantiate(getConf());
-2140UserGroupInformation ugi =
userProvider.getCurrent().getUGI();
-2141FileStatus[] files =
fs.listStatus(hbaseDir);
-2142for (FileStatus file : files) {
-2143 try {
-2144FSUtils.checkAccess(ugi, file,
FsAction.WRITE);
-2145 } catch (AccessDeniedException
ace) {
-2146LOG.warn("Got
AccessDeniedException when preCheckPermission ", ace);
-2147
errors.reportError(ERROR_CODE.WRONG_USAGE, "Current user " +
ugi.getUserName()
-2148 + " does not have write perms
to " + file.getPath()
-2149 + ". Please rerun hbck as hdfs
user " + file.getOwner());
-2150throw ace;
-2151 }
-2152}
-2153 }
-2154
-2155 /**
-2156 * Deletes region from meta table
-2157 */
-2158 private void deleteMetaRegion(HbckInfo
hi) throws IOException {
-2159
deleteMetaRegion(hi.metaEntry.getRegionName());
-2160 }
-2161
-2162 /**
-2163 * Deletes region from meta table
-2164 */
-2165 private void deleteMetaRegion(byte[]
metaKey) throws IOException {
-2166Delete d = new Delete(metaKey);
-2167meta.delete(d);
-2168LOG.info("Deleted " +
Bytes.toString(metaKey) + " from META" );
-2169 }
-2170
-2171 /**
-2172 * Reset the split parent region info
in meta table
-2173 */
-2174 private void resetSplitParent(HbckInfo
hi) throws IOException {
-2175RowMutations mutations = new
RowMutations(hi.metaEntry.getRegionName());
-2176Delete d = new
Delete(hi.metaEntry.getRegionName());
-2177
d.addColumn(HConstants.CATALOG_FAMILY, HConstants.SPLITA_QUALIFIER);
-2178
d.addColumn(HConstants.CATALOG_FAMILY, HConstants.SPLITB_QUALIFIER);
-2179mutations.add(d);
-2180
-2181RegionInfo hri =
RegionInfoBuilder.newBuilder(hi.metaEntry)
-2182.setOffline(false)
-2183.setSplit(false)
-2184.build();
-2185Put p =
MetaTableAccessor.makePutFromRegionInfo(hri,
EnvironmentEdgeManager.currentTime());
-2186mutations.add(p);
-2187
-2188meta.mutateRow(mutations);
-2189LOG.info("Reset split parent " +
hi.metaEntry.getRegionNameAsString() + " in META" );
-2190 }
-2191
-2192 /**
-2193 * This backwards-compatibility
wrapper for permanently offlining a region
-2194 * that should not be alive. If the
region server does not support the
-2195 * "offline" method, it will use the
closest unassign method instead. This
-2196 * will basically work until one
attempts to disable or delete the affected
-2197 * table. The problem has to do with
in-memory only master state, so
-2198 * restarting the HMaster or failing
over to another should fix this.
-2199 */
-2200 private void offline(byte[]
regionName) throws IOException {
-2201String regionString =
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e0fb1fde/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
index 79bf967..c8b113b 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.RedirectServlet.html
@@ -115,3514 +115,3517 @@
107import
org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer;
108import
org.apache.hadoop.hbase.master.balancer.ClusterStatusChore;
109import
org.apache.hadoop.hbase.master.balancer.LoadBalancerFactory;
-110import
org.apache.hadoop.hbase.master.cleaner.HFileCleaner;
-111import
org.apache.hadoop.hbase.master.cleaner.LogCleaner;
-112import
org.apache.hadoop.hbase.master.cleaner.ReplicationBarrierCleaner;
-113import
org.apache.hadoop.hbase.master.locking.LockManager;
-114import
org.apache.hadoop.hbase.master.normalizer.NormalizationPlan;
-115import
org.apache.hadoop.hbase.master.normalizer.NormalizationPlan.PlanType;
-116import
org.apache.hadoop.hbase.master.normalizer.RegionNormalizer;
-117import
org.apache.hadoop.hbase.master.normalizer.RegionNormalizerChore;
-118import
org.apache.hadoop.hbase.master.normalizer.RegionNormalizerFactory;
-119import
org.apache.hadoop.hbase.master.procedure.CreateTableProcedure;
-120import
org.apache.hadoop.hbase.master.procedure.DeleteNamespaceProcedure;
-121import
org.apache.hadoop.hbase.master.procedure.DeleteTableProcedure;
-122import
org.apache.hadoop.hbase.master.procedure.DisableTableProcedure;
-123import
org.apache.hadoop.hbase.master.procedure.EnableTableProcedure;
-124import
org.apache.hadoop.hbase.master.procedure.MasterProcedureConstants;
-125import
org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
-126import
org.apache.hadoop.hbase.master.procedure.MasterProcedureScheduler;
-127import
org.apache.hadoop.hbase.master.procedure.MasterProcedureUtil;
-128import
org.apache.hadoop.hbase.master.procedure.ModifyTableProcedure;
-129import
org.apache.hadoop.hbase.master.procedure.ProcedurePrepareLatch;
-130import
org.apache.hadoop.hbase.master.procedure.RecoverMetaProcedure;
-131import
org.apache.hadoop.hbase.master.procedure.TruncateTableProcedure;
-132import
org.apache.hadoop.hbase.master.replication.AddPeerProcedure;
-133import
org.apache.hadoop.hbase.master.replication.DisablePeerProcedure;
-134import
org.apache.hadoop.hbase.master.replication.EnablePeerProcedure;
-135import
org.apache.hadoop.hbase.master.replication.ModifyPeerProcedure;
-136import
org.apache.hadoop.hbase.master.replication.RemovePeerProcedure;
-137import
org.apache.hadoop.hbase.master.replication.ReplicationPeerManager;
-138import
org.apache.hadoop.hbase.master.replication.UpdatePeerConfigProcedure;
-139import
org.apache.hadoop.hbase.master.snapshot.SnapshotManager;
-140import
org.apache.hadoop.hbase.mob.MobConstants;
-141import
org.apache.hadoop.hbase.monitoring.MemoryBoundedLogMessageBuffer;
-142import
org.apache.hadoop.hbase.monitoring.MonitoredTask;
-143import
org.apache.hadoop.hbase.monitoring.TaskMonitor;
-144import
org.apache.hadoop.hbase.procedure.MasterProcedureManagerHost;
-145import
org.apache.hadoop.hbase.procedure.flush.MasterFlushTableProcedureManager;
-146import
org.apache.hadoop.hbase.procedure2.LockedResource;
-147import
org.apache.hadoop.hbase.procedure2.Procedure;
-148import
org.apache.hadoop.hbase.procedure2.ProcedureEvent;
-149import
org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
-150import
org.apache.hadoop.hbase.procedure2.RemoteProcedureDispatcher.RemoteProcedure;
-151import
org.apache.hadoop.hbase.procedure2.RemoteProcedureException;
-152import
org.apache.hadoop.hbase.procedure2.store.wal.WALProcedureStore;
-153import
org.apache.hadoop.hbase.quotas.MasterQuotaManager;
-154import
org.apache.hadoop.hbase.quotas.MasterSpaceQuotaObserver;
-155import
org.apache.hadoop.hbase.quotas.QuotaObserverChore;
-156import
org.apache.hadoop.hbase.quotas.QuotaUtil;
-157import
org.apache.hadoop.hbase.quotas.SnapshotQuotaObserverChore;
-158import
org.apache.hadoop.hbase.quotas.SpaceQuotaSnapshotNotifier;
-159import
org.apache.hadoop.hbase.quotas.SpaceQuotaSnapshotNotifierFactory;
-160import
org.apache.hadoop.hbase.regionserver.DefaultStoreEngine;
-161import
org.apache.hadoop.hbase.regionserver.HRegionServer;
-162import
org.apache.hadoop.hbase.regionserver.HStore;
-163import
org.apache.hadoop.hbase.regionserver.RSRpcServices;
-164import
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost;
-165import
org.apache.hadoop.hbase.regionserver.RegionSplitPolicy;
-166import
org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy;
-167import
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/16541468/devapidocs/org/apache/hadoop/hbase/client/HTable.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/HTable.html
b/devapidocs/org/apache/hadoop/hbase/client/HTable.html
index 44fcf94..dd6854b 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/HTable.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/HTable.html
@@ -115,7 +115,7 @@ var activeTableTab = "activeTableTab";
@InterfaceAudience.Private
@InterfaceStability.Stable
-public class HTable
+public class HTable
extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
implements Table
An implementation of Table.
Used to communicate with a single HBase table.
@@ -515,29 +515,32 @@ implements
private boolean
-doCheckAndDelete(byte[]row,
+doCheckAndDelete(byte[]row,
byte[]family,
byte[]qualifier,
https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">StringopName,
byte[]value,
+TimeRangetimeRange,
Deletedelete)
private boolean
-doCheckAndMutate(byte[]row,
+doCheckAndMutate(byte[]row,
byte[]family,
byte[]qualifier,
https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">StringopName,
byte[]value,
+TimeRangetimeRange,
RowMutationsrm)
private boolean
-doCheckAndPut(byte[]row,
+doCheckAndPut(byte[]row,
byte[]family,
byte[]qualifier,
https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">StringopName,
byte[]value,
+ TimeRangetimeRange,
Putput)
@@ -832,7 +835,7 @@ implements
LOG
-private static finalorg.slf4j.Logger LOG
+private static finalorg.slf4j.Logger LOG
@@ -841,7 +844,7 @@ implements
DEFAULT_CONSISTENCY
-private static finalConsistency DEFAULT_CONSISTENCY
+private static finalConsistency DEFAULT_CONSISTENCY
@@ -850,7 +853,7 @@ implements
connection
-private finalClusterConnection connection
+private finalClusterConnection connection
@@ -859,7 +862,7 @@ implements
tableName
-private finalTableName tableName
+private finalTableName tableName
@@ -868,7 +871,7 @@ implements
configuration
-private finalorg.apache.hadoop.conf.Configuration configuration
+private finalorg.apache.hadoop.conf.Configuration configuration
@@ -877,7 +880,7 @@ implements
connConfiguration
-private finalConnectionConfiguration connConfiguration
+private finalConnectionConfiguration connConfiguration
@@ -886,7 +889,7 @@ implements
closed
-privateboolean closed
+privateboolean closed
@@ -895,7 +898,7 @@ implements
scannerCaching
-private finalint scannerCaching
+private finalint scannerCaching
@@ -904,7 +907,7 @@ implements
scannerMaxResultSize
-private finallong scannerMaxResultSize
+private finallong scannerMaxResultSize
@@ -913,7 +916,7 @@ implements
pool
-private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true;
title="class or interface in java.util.concurrent">ExecutorService pool
+private finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true;
title="class or interface in java.util.concurrent">ExecutorService pool
@@ -922,7 +925,7 @@ implements
operationTimeoutMs
-privateint operationTimeoutMs
+privateint operationTimeoutMs
@@ -931,7 +934,7 @@ implements
rpcTimeoutMs
-private finalint rpcTimeoutMs
+private finalint rpcTimeoutMs
@@ -940,7 +943,7 @@ implements
readRpcTimeoutMs
-privateint readRpcTimeoutMs
+privateint readRpcTimeoutMs
@@ -949,7 +952,7 @@ implements
writeRpcTimeoutMs
-privateint writeRpcTimeoutMs
+privateint writeRpcTimeoutMs
@@ -958,7 +961,7 @@ implements
cleanupPoolOnClose
-private finalboolean cleanupPoolOnClose
+private finalboolean cleanupPoolOnClose
@@ -967,7 +970,7 @@ implements
locator
-private finalHRegionLocator locator
+private finalHRegionLocator locator
@@ -976,7 +979,7 @@ implements
multiAp
-AsyncProcess multiAp
+AsyncProcess multiAp
The Async process for batch
@@ -986,7 +989,7 @@ implements
rpcCallerFactory
-private finalRpcRetryingCallerFactory
rpcCallerFactory
+private finalRpcRetryingCallerFactory
rpcCallerFactory
@@ -995,7 +998,7 @@ implements
rpcControllerFactory
-private finalRpcControllerFactory rpcControllerFactory
+private finalRpcControllerFactory rpcControllerFactory
@@ -1013,7 +1016,7 @@ implements
HTable
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f1ebf5b6/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html
b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html
index de426ec..c0b3f3f 100644
--- a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html
+++ b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html
@@ -168,39 +168,27 @@
void
-CombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
- Cacheablebuf)
-
-
-void
BlockCache.cacheBlock(BlockCacheKeycacheKey,
Cacheablebuf)
Add block to cache (defaults to not in-memory).
-
+
void
LruBlockCache.cacheBlock(BlockCacheKeycacheKey,
Cacheablebuf)
Cache the block with the specified name and buffer.
-
-void
-MemcachedBlockCache.cacheBlock(BlockCacheKeycacheKey,
- Cacheablebuf)
-
void
-CombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
- Cacheablebuf,
- booleaninMemory)
+CombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
+ Cacheablebuf)
void
-InclusiveCombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
- Cacheablebuf,
- booleaninMemory)
+MemcachedBlockCache.cacheBlock(BlockCacheKeycacheKey,
+ Cacheablebuf)
void
@@ -220,6 +208,18 @@
void
+CombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
+ Cacheablebuf,
+ booleaninMemory)
+
+
+void
+InclusiveCombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
+ Cacheablebuf,
+ booleaninMemory)
+
+
+void
MemcachedBlockCache.cacheBlock(BlockCacheKeycacheKey,
Cacheablebuf,
booleaninMemory)
@@ -232,21 +232,21 @@
boolean
-CombinedBlockCache.evictBlock(BlockCacheKeycacheKey)
+BlockCache.evictBlock(BlockCacheKeycacheKey)
+Evict block from cache.
+
boolean
-InclusiveCombinedBlockCache.evictBlock(BlockCacheKeycacheKey)
+LruBlockCache.evictBlock(BlockCacheKeycacheKey)
boolean
-BlockCache.evictBlock(BlockCacheKeycacheKey)
-Evict block from cache.
-
+CombinedBlockCache.evictBlock(BlockCacheKeycacheKey)
boolean
-LruBlockCache.evictBlock(BlockCacheKeycacheKey)
+InclusiveCombinedBlockCache.evictBlock(BlockCacheKeycacheKey)
boolean
@@ -254,35 +254,35 @@
Cacheable
-CombinedBlockCache.getBlock(BlockCacheKeycacheKey,
+BlockCache.getBlock(BlockCacheKeycacheKey,
booleancaching,
booleanrepeat,
-booleanupdateCacheMetrics)
+booleanupdateCacheMetrics)
+Fetch block from cache.
+
Cacheable
-InclusiveCombinedBlockCache.getBlock(BlockCacheKeycacheKey,
+LruBlockCache.getBlock(BlockCacheKeycacheKey,
booleancaching,
booleanrepeat,
-booleanupdateCacheMetrics)
+booleanupdateCacheMetrics)
+Get the buffer of the block with the specified name.
+
Cacheable
-BlockCache.getBlock(BlockCacheKeycacheKey,
+CombinedBlockCache.getBlock(BlockCacheKeycacheKey,
booleancaching,
booleanrepeat,
-booleanupdateCacheMetrics)
-Fetch block from cache.
-
+booleanupdateCacheMetrics)
Cacheable
-LruBlockCache.getBlock(BlockCacheKeycacheKey,
+InclusiveCombinedBlockCache.getBlock(BlockCacheKeycacheKey,
booleancaching,
booleanrepeat,
-booleanupdateCacheMetrics)
-Get the buffer of the block with the specified name.
-
+booleanupdateCacheMetrics)
Cacheable
@@ -308,11 +308,6 @@
CombinedBlockCache.getRefCount(BlockCacheKeycacheKey)
-void
-CombinedBlockCache.returnBlock(BlockCacheKeycacheKey,
- Cacheableblock)
-
-
default void
BlockCache.returnBlock(BlockCacheKeycacheKey,
Cacheableblock)
@@ -320,6 +315,11 @@
is over.
+
+void
+CombinedBlockCache.returnBlock(BlockCacheKeycacheKey,
+ Cacheableblock)
+
@@ -497,14 +497,14 @@
void
-BucketCache.BucketEntryGroup.add(https://docs.oracle.com/javase/8/docs/api/java/util/Map.Entry.html?is-external=true;
title="class or interface in java.util">Map.EntryBlockCacheKey,BucketCache.BucketEntryblock)
-
-
-void
CachedEntryQueue.add(https://docs.oracle.com/javase/8/docs/api/java/util/Map.Entry.html?is-external=true;
title="class or interface in java.util">Map.EntryBlockCacheKey,BucketCache.BucketEntryentry)
Attempt to add the specified entry to this queue.
+
+void
+BucketCache.BucketEntryGroup.add(https://docs.oracle.com/javase/8/docs/api/java/util/Map.Entry.html?is-external=true;
title="class or interface in java.util">Map.EntryBlockCacheKey,BucketCache.BucketEntryblock)
+
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f1ebf5b6/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockType.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockType.html
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4dc2a2e8/devapidocs/org/apache/hadoop/hbase/filter/class-use/Filter.ReturnCode.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/filter/class-use/Filter.ReturnCode.html
b/devapidocs/org/apache/hadoop/hbase/filter/class-use/Filter.ReturnCode.html
index 3b7627b..5f975db 100644
--- a/devapidocs/org/apache/hadoop/hbase/filter/class-use/Filter.ReturnCode.html
+++ b/devapidocs/org/apache/hadoop/hbase/filter/class-use/Filter.ReturnCode.html
@@ -151,119 +151,119 @@
Filter.ReturnCode
-FilterListWithAND.filterCell(Cellc)
+ColumnPrefixFilter.filterCell(Cellcell)
Filter.ReturnCode
-ValueFilter.filterCell(Cellc)
+ColumnCountGetFilter.filterCell(Cellc)
Filter.ReturnCode
-SkipFilter.filterCell(Cellc)
+RowFilter.filterCell(Cellv)
Filter.ReturnCode
-FamilyFilter.filterCell(Cellc)
+FuzzyRowFilter.filterCell(Cellc)
Filter.ReturnCode
-ColumnPrefixFilter.filterCell(Cellcell)
+Filter.filterCell(Cellc)
+A way to filter based on the column family, column
qualifier and/or the column value.
+
Filter.ReturnCode
-PageFilter.filterCell(Cellignored)
+RandomRowFilter.filterCell(Cellc)
Filter.ReturnCode
-RowFilter.filterCell(Cellv)
+FirstKeyOnlyFilter.filterCell(Cellc)
Filter.ReturnCode
-ColumnRangeFilter.filterCell(Cellc)
+SkipFilter.filterCell(Cellc)
Filter.ReturnCode
-ColumnCountGetFilter.filterCell(Cellc)
+TimestampsFilter.filterCell(Cellc)
Filter.ReturnCode
-MultipleColumnPrefixFilter.filterCell(Cellc)
+ValueFilter.filterCell(Cellc)
Filter.ReturnCode
-ColumnPaginationFilter.filterCell(Cellc)
+KeyOnlyFilter.filterCell(Cellignored)
Filter.ReturnCode
-DependentColumnFilter.filterCell(Cellc)
+FamilyFilter.filterCell(Cellc)
Filter.ReturnCode
-FilterListWithOR.filterCell(Cellc)
+QualifierFilter.filterCell(Cellc)
Filter.ReturnCode
-InclusiveStopFilter.filterCell(Cellc)
+FilterList.filterCell(Cellc)
Filter.ReturnCode
-KeyOnlyFilter.filterCell(Cellignored)
+ColumnRangeFilter.filterCell(Cellc)
Filter.ReturnCode
-MultiRowRangeFilter.filterCell(Cellignored)
+ColumnPaginationFilter.filterCell(Cellc)
Filter.ReturnCode
-Filter.filterCell(Cellc)
-A way to filter based on the column family, column
qualifier and/or the column value.
-
+FilterListWithAND.filterCell(Cellc)
Filter.ReturnCode
-FirstKeyOnlyFilter.filterCell(Cellc)
+WhileMatchFilter.filterCell(Cellc)
Filter.ReturnCode
-WhileMatchFilter.filterCell(Cellc)
+ColumnValueFilter.filterCell(Cellc)
Filter.ReturnCode
-FirstKeyValueMatchingQualifiersFilter.filterCell(Cellc)
-Deprecated.
-
+MultiRowRangeFilter.filterCell(Cellignored)
Filter.ReturnCode
-TimestampsFilter.filterCell(Cellc)
+PrefixFilter.filterCell(Cellc)
Filter.ReturnCode
-FuzzyRowFilter.filterCell(Cellc)
+DependentColumnFilter.filterCell(Cellc)
Filter.ReturnCode
-FilterList.filterCell(Cellc)
+FirstKeyValueMatchingQualifiersFilter.filterCell(Cellc)
+Deprecated.
+
Filter.ReturnCode
-RandomRowFilter.filterCell(Cellc)
+PageFilter.filterCell(Cellignored)
Filter.ReturnCode
-PrefixFilter.filterCell(Cellc)
+FilterListWithOR.filterCell(Cellc)
Filter.ReturnCode
-SingleColumnValueFilter.filterCell(Cellc)
+InclusiveStopFilter.filterCell(Cellc)
Filter.ReturnCode
-ColumnValueFilter.filterCell(Cellc)
+MultipleColumnPrefixFilter.filterCell(Cellc)
Filter.ReturnCode
-QualifierFilter.filterCell(Cellc)
+SingleColumnValueFilter.filterCell(Cellc)
Filter.ReturnCode
@@ -279,158 +279,158 @@
Filter.ReturnCode
-ValueFilter.filterKeyValue(Cellc)
+ColumnPrefixFilter.filterKeyValue(Cellc)
Deprecated.
Filter.ReturnCode
-SkipFilter.filterKeyValue(Cellc)
+ColumnCountGetFilter.filterKeyValue(Cellc)
Deprecated.
Filter.ReturnCode
-FilterListBase.filterKeyValue(Cellc)
+RowFilter.filterKeyValue(Cellc)
+Deprecated.
+
Filter.ReturnCode
-FamilyFilter.filterKeyValue(Cellc)
+FuzzyRowFilter.filterKeyValue(Cellc)
Deprecated.
Filter.ReturnCode
-ColumnPrefixFilter.filterKeyValue(Cellc)
-Deprecated.
+Filter.filterKeyValue(Cellc)
+Deprecated.
+As of release 2.0.0, this
will be removed in HBase 3.0.0.
+ Instead use filterCell(Cell)
+
Filter.ReturnCode
-PageFilter.filterKeyValue(Cellc)
+RandomRowFilter.filterKeyValue(Cellc)
Deprecated.
Filter.ReturnCode
-RowFilter.filterKeyValue(Cellc)
+FirstKeyOnlyFilter.filterKeyValue(Cellc)
Deprecated.
Filter.ReturnCode
-ColumnRangeFilter.filterKeyValue(Cellc)
+SkipFilter.filterKeyValue(Cellc)
Deprecated.
Filter.ReturnCode
-ColumnCountGetFilter.filterKeyValue(Cellc)
+TimestampsFilter.filterKeyValue(Cellc)
Deprecated.
Filter.ReturnCode
-MultipleColumnPrefixFilter.filterKeyValue(Cellc)
+ValueFilter.filterKeyValue(Cellc)
Deprecated.
Filter.ReturnCode
-ColumnPaginationFilter.filterKeyValue(Cellc)
+KeyOnlyFilter.filterKeyValue(Cellignored)
Deprecated.
Filter.ReturnCode
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6eb695c8/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html
index ce22ca5..64099d9 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/ProtobufLogWriter.html
@@ -80,50 +80,57 @@
072 }
073
074 @Override
-075 public void sync() throws IOException
{
+075 public void sync(boolean forceSync)
throws IOException {
076FSDataOutputStream fsdos =
this.output;
077if (fsdos == null) return; // Presume
closed
078fsdos.flush();
-079fsdos.hflush();
-080 }
-081
-082 public FSDataOutputStream getStream()
{
-083return this.output;
+079if (forceSync) {
+080 fsdos.hsync();
+081} else {
+082 fsdos.hflush();
+083}
084 }
085
-086 @SuppressWarnings("deprecation")
-087 @Override
-088 protected void initOutput(FileSystem
fs, Path path, boolean overwritable, int bufferSize,
-089 short replication, long blockSize)
throws IOException, StreamLacksCapabilityException {
-090this.output =
fs.createNonRecursive(path, overwritable, bufferSize, replication, blockSize,
-091 null);
-092// TODO Be sure to add a check for
hsync if this branch includes HBASE-19024
-093if
(fs.getConf().getBoolean(CommonFSUtils.UNSAFE_STREAM_CAPABILITY_ENFORCE, true)
-094
!(CommonFSUtils.hasCapability(output, "hflush"))) {
-095 throw new
StreamLacksCapabilityException("hflush");
-096}
-097 }
-098
-099 @Override
-100 protected long
writeMagicAndWALHeader(byte[] magic, WALHeader header) throws IOException {
-101output.write(magic);
-102header.writeDelimitedTo(output);
-103return output.getPos();
+086 public FSDataOutputStream getStream()
{
+087return this.output;
+088 }
+089
+090 @SuppressWarnings("deprecation")
+091 @Override
+092 protected void initOutput(FileSystem
fs, Path path, boolean overwritable, int bufferSize,
+093 short replication, long blockSize)
throws IOException, StreamLacksCapabilityException {
+094this.output =
fs.createNonRecursive(path, overwritable, bufferSize, replication, blockSize,
+095 null);
+096if
(fs.getConf().getBoolean(CommonFSUtils.UNSAFE_STREAM_CAPABILITY_ENFORCE, true))
{
+097 if
(!CommonFSUtils.hasCapability(output, "hflush")) {
+098throw new
StreamLacksCapabilityException("hflush");
+099 }
+100 if
(!CommonFSUtils.hasCapability(output, "hsync")) {
+101throw new
StreamLacksCapabilityException("hsync");
+102 }
+103}
104 }
105
106 @Override
-107 protected OutputStream
getOutputStreamForCellEncoder() {
-108return this.output;
-109 }
-110
-111 @Override
-112 protected long
writeWALTrailerAndMagic(WALTrailer trailer, byte[] magic) throws IOException
{
-113trailer.writeTo(output);
-114
output.writeInt(trailer.getSerializedSize());
-115output.write(magic);
-116return output.getPos();
-117 }
-118}
+107 protected long
writeMagicAndWALHeader(byte[] magic, WALHeader header) throws IOException {
+108output.write(magic);
+109header.writeDelimitedTo(output);
+110return output.getPos();
+111 }
+112
+113 @Override
+114 protected OutputStream
getOutputStreamForCellEncoder() {
+115return this.output;
+116 }
+117
+118 @Override
+119 protected long
writeWALTrailerAndMagic(WALTrailer trailer, byte[] magic) throws IOException
{
+120trailer.writeTo(output);
+121
output.writeInt(trailer.getSerializedSize());
+122output.write(magic);
+123return output.getPos();
+124 }
+125}
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6eb695c8/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/SyncFuture.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/SyncFuture.html
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/SyncFuture.html
index 9457b4b..6a84f26 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/SyncFuture.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/wal/SyncFuture.html
@@ -75,94 +75,105 @@
067
068 private Thread t;
069
-070 /**
-071 * Call this method to clear old usage
and get it ready for new deploy.
-072 * @param txid the new transaction id
-073 * @param span current span, detached
from caller. Don't forget to attach it when resuming after a
-074 * call to {@link
#get(long)}.
-075 * @return this
-076 */
-077 synchronized SyncFuture reset(long
txid) {
-078if (t != null t !=
Thread.currentThread()) {
-079
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bd675fa3/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html
index 3bc66bb..97aa79c 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.KeepAliveWorkerThread.html
@@ -1435,459 +1435,460 @@
1427 */
1428 private void execProcedure(final
RootProcedureState procStack,
1429 final
ProcedureTEnvironment procedure) {
-1430
Preconditions.checkArgument(procedure.getState() == ProcedureState.RUNNABLE);
-1431
-1432// Procedures can suspend
themselves. They skip out by throwing a ProcedureSuspendedException.
-1433// The exception is caught below and
then we hurry to the exit without disturbing state. The
-1434// idea is that the processing of
this procedure will be unsuspended later by an external event
-1435// such the report of a region open.
TODO: Currently, its possible for two worker threads
-1436// to be working on the same
procedure concurrently (locking in procedures is NOT about
-1437// concurrency but about tying an
entity to a procedure; i.e. a region to a particular
-1438// procedure instance). This can
make for issues if both threads are changing state.
-1439// See
env.getProcedureScheduler().wakeEvent(regionNode.getProcedureEvent());
-1440// in
RegionTransitionProcedure#reportTransition for example of Procedure putting
-1441// itself back on the scheduler
making it possible for two threads running against
-1442// the one Procedure. Might be ok if
they are both doing different, idempotent sections.
-1443boolean suspended = false;
-1444
-1445// Whether to 're-' -execute; run
through the loop again.
-1446boolean reExecute = false;
-1447
-1448ProcedureTEnvironment[]
subprocs = null;
-1449do {
-1450 reExecute = false;
-1451 try {
-1452subprocs =
procedure.doExecute(getEnvironment());
-1453if (subprocs != null
subprocs.length == 0) {
-1454 subprocs = null;
-1455}
-1456 } catch
(ProcedureSuspendedException e) {
-1457if (LOG.isTraceEnabled()) {
-1458 LOG.trace("Suspend " +
procedure);
-1459}
-1460suspended = true;
-1461 } catch (ProcedureYieldException
e) {
-1462if (LOG.isTraceEnabled()) {
-1463 LOG.trace("Yield " + procedure
+ ": " + e.getMessage(), e);
-1464}
-1465scheduler.yield(procedure);
-1466return;
-1467 } catch (InterruptedException e)
{
-1468if (LOG.isTraceEnabled()) {
-1469 LOG.trace("Yield interrupt " +
procedure + ": " + e.getMessage(), e);
-1470}
-1471
handleInterruptedException(procedure, e);
-1472scheduler.yield(procedure);
-1473return;
-1474 } catch (Throwable e) {
-1475// Catch NullPointerExceptions
or similar errors...
-1476String msg = "CODE-BUG: Uncaught
runtime exception: " + procedure;
-1477LOG.error(msg, e);
-1478procedure.setFailure(new
RemoteProcedureException(msg, e));
-1479 }
-1480
-1481 if (!procedure.isFailed()) {
-1482if (subprocs != null) {
-1483 if (subprocs.length == 1
subprocs[0] == procedure) {
-1484// Procedure returned
itself. Quick-shortcut for a state machine-like procedure;
-1485// i.e. we go around this
loop again rather than go back out on the scheduler queue.
-1486subprocs = null;
-1487reExecute = true;
-1488if (LOG.isTraceEnabled())
{
-1489 LOG.trace("Short-circuit
to next step on pid=" + procedure.getProcId());
-1490}
-1491 } else {
-1492// Yield the current
procedure, and make the subprocedure runnable
-1493// subprocs may come back
'null'.
-1494subprocs =
initializeChildren(procStack, procedure, subprocs);
-1495LOG.info("Initialized
subprocedures=" +
-1496 (subprocs == null? null:
-1497
Stream.of(subprocs).map(e - "{" + e.toString() + "}").
-1498
collect(Collectors.toList()).toString()));
-1499 }
-1500} else if (procedure.getState()
== ProcedureState.WAITING_TIMEOUT) {
-1501 if (LOG.isTraceEnabled()) {
-1502LOG.trace("Added to
timeoutExecutor " + procedure);
-1503 }
-1504
timeoutExecutor.add(procedure);
-1505} else if (!suspended) {
-1506 // No subtask, so we are
done
-1507
procedure.setState(ProcedureState.SUCCESS);
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8b1eaec1/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
--
diff --git
a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
index 3a91ea6..670bcf3 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/class-use/HBaseClassTestRule.html
@@ -975,630 +975,634 @@
static HBaseClassTestRule
-TestImmutableHRegionInfo.CLASS_RULE
+TestCoprocessorDescriptor.CLASS_RULE
static HBaseClassTestRule
-TestMutation.CLASS_RULE
+TestImmutableHRegionInfo.CLASS_RULE
static HBaseClassTestRule
-TestRowComparator.CLASS_RULE
+TestMutation.CLASS_RULE
static HBaseClassTestRule
-TestBufferedMutator.CLASS_RULE
+TestRowComparator.CLASS_RULE
static HBaseClassTestRule
-TestClientScanner.CLASS_RULE
+TestBufferedMutator.CLASS_RULE
static HBaseClassTestRule
-TestInterfaceAlign.CLASS_RULE
+TestClientScanner.CLASS_RULE
static HBaseClassTestRule
-TestBufferedMutatorParams.CLASS_RULE
+TestInterfaceAlign.CLASS_RULE
static HBaseClassTestRule
-TestSnapshotFromAdmin.CLASS_RULE
+TestBufferedMutatorParams.CLASS_RULE
static HBaseClassTestRule
-TestIncrement.CLASS_RULE
+TestSnapshotFromAdmin.CLASS_RULE
static HBaseClassTestRule
-TestAsyncProcessWithRegionException.CLASS_RULE
+TestIncrement.CLASS_RULE
static HBaseClassTestRule
-TestHTableMultiplexerViaMocks.CLASS_RULE
+TestAsyncProcessWithRegionException.CLASS_RULE
static HBaseClassTestRule
-TestColumnFamilyDescriptorBuilder.CLASS_RULE
+TestHTableMultiplexerViaMocks.CLASS_RULE
static HBaseClassTestRule
-TestMetricsConnection.CLASS_RULE
+TestColumnFamilyDescriptorBuilder.CLASS_RULE
static HBaseClassTestRule
-TestTableDescriptorBuilder.CLASS_RULE
+TestMetricsConnection.CLASS_RULE
static HBaseClassTestRule
-TestImmutableHColumnDescriptor.CLASS_RULE
+TestTableDescriptorBuilder.CLASS_RULE
static HBaseClassTestRule
-TestImmutableHTableDescriptor.CLASS_RULE
+TestImmutableHColumnDescriptor.CLASS_RULE
static HBaseClassTestRule
-TestRetriesExhaustedWithDetailsException.CLASS_RULE
+TestImmutableHTableDescriptor.CLASS_RULE
static HBaseClassTestRule
-TestAsyncProcess.CLASS_RULE
+TestRetriesExhaustedWithDetailsException.CLASS_RULE
static HBaseClassTestRule
-TestProcedureFuture.CLASS_RULE
+TestAsyncProcess.CLASS_RULE
static HBaseClassTestRule
-TestClientExponentialBackoff.CLASS_RULE
+TestProcedureFuture.CLASS_RULE
static HBaseClassTestRule
-TestDeleteTimeStamp.CLASS_RULE
+TestClientExponentialBackoff.CLASS_RULE
static HBaseClassTestRule
-TestOperation.CLASS_RULE
+TestDeleteTimeStamp.CLASS_RULE
static HBaseClassTestRule
-TestMultiParallel.CLASS_RULE
+TestOperation.CLASS_RULE
static HBaseClassTestRule
-TestSnapshotMetadata.CLASS_RULE
+TestMultiParallel.CLASS_RULE
static HBaseClassTestRule
-TestFromClientSideNoCodec.CLASS_RULE
+TestSnapshotMetadata.CLASS_RULE
static HBaseClassTestRule
-TestAsyncBufferMutator.CLASS_RULE
+TestFromClientSideNoCodec.CLASS_RULE
static HBaseClassTestRule
-TestSizeFailures.CLASS_RULE
+TestAsyncBufferMutator.CLASS_RULE
static HBaseClassTestRule
-TestGetProcedureResult.CLASS_RULE
+TestSizeFailures.CLASS_RULE
static HBaseClassTestRule
-TestAsyncTableScanRenewLease.CLASS_RULE
+TestGetProcedureResult.CLASS_RULE
static HBaseClassTestRule
-TestClientOperationInterrupt.CLASS_RULE
+TestAsyncTableScanRenewLease.CLASS_RULE
static HBaseClassTestRule
-TestScanWithoutFetchingData.CLASS_RULE
+TestClientOperationInterrupt.CLASS_RULE
static HBaseClassTestRule
-TestServerLoadDurability.CLASS_RULE
+TestScanWithoutFetchingData.CLASS_RULE
static HBaseClassTestRule
-TestAsyncTableScan.CLASS_RULE
+TestServerLoadDurability.CLASS_RULE
static HBaseClassTestRule
-TestMalformedCellFromClient.CLASS_RULE
+TestAsyncTableScan.CLASS_RULE
static HBaseClassTestRule
-TestAsyncReplicationAdminApiWithClusters.CLASS_RULE
+TestMalformedCellFromClient.CLASS_RULE
static HBaseClassTestRule
-TestCheckAndMutate.CLASS_RULE
+TestAsyncReplicationAdminApiWithClusters.CLASS_RULE
static HBaseClassTestRule
-TestFastFail.CLASS_RULE
+TestCheckAndMutate.CLASS_RULE
static HBaseClassTestRule
-TestAppendFromClientSide.CLASS_RULE
+TestFastFail.CLASS_RULE
static HBaseClassTestRule
-TestServerBusyException.CLASS_RULE
+TestAppendFromClientSide.CLASS_RULE
static HBaseClassTestRule
-TestMobSnapshotFromClient.CLASS_RULE
+TestServerBusyException.CLASS_RULE
static HBaseClassTestRule
-TestAsyncQuotaAdminApi.CLASS_RULE
+TestMobSnapshotFromClient.CLASS_RULE
static HBaseClassTestRule
-TestSplitOrMergeStatus.CLASS_RULE
+TestAsyncQuotaAdminApi.CLASS_RULE
static HBaseClassTestRule
-TestFromClientSideScanExcpetionWithCoprocessor.CLASS_RULE
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/a754d895/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html
b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html
index ec5a869..84f7079 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/RSRpcServices.html
@@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab";
@InterfaceAudience.Private
-public class RSRpcServices
+public class RSRpcServices
extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
implements HBaseRPCErrorHandler,
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService.BlockingInterface,
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ClientService.BlockingInterface,
PriorityFunction, ConfigurationObserver
Implements the regionserver RPC services.
@@ -191,7 +191,7 @@ implements Field and Description
-private AccessChecker
+protected AccessChecker
accessChecker
@@ -959,7 +959,7 @@ implements
LOG
-protected static finalorg.slf4j.Logger LOG
+protected static finalorg.slf4j.Logger LOG
@@ -968,7 +968,7 @@ implements
REGION_SERVER_RPC_SCHEDULER_FACTORY_CLASS
-public static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String REGION_SERVER_RPC_SCHEDULER_FACTORY_CLASS
+public static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String REGION_SERVER_RPC_SCHEDULER_FACTORY_CLASS
RPC scheduler to use for the region server.
See Also:
@@ -982,7 +982,7 @@ implements
REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
-private static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
+private static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
Minimum allowable time limit delta (in milliseconds) that
can be enforced during scans. This
configuration exists to prevent the scenario where a time limit is specified
to be so
restrictive that the time limit is reached immediately (before any cells are
scanned).
@@ -998,7 +998,7 @@ implements
DEFAULT_REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
-private static finallong DEFAULT_REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
+private static finallong DEFAULT_REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
Default value of REGION_SERVER_RPC_MINIMUM_SCAN_TIME_LIMIT_DELTA
See Also:
@@ -1012,7 +1012,7 @@ implements
BATCH_ROWS_THRESHOLD_NAME
-static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String BATCH_ROWS_THRESHOLD_NAME
+static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String BATCH_ROWS_THRESHOLD_NAME
Number of rows in a batch operation above which a warning
will be logged.
See Also:
@@ -1026,7 +1026,7 @@ implements
BATCH_ROWS_THRESHOLD_DEFAULT
-static finalint BATCH_ROWS_THRESHOLD_DEFAULT
+static finalint BATCH_ROWS_THRESHOLD_DEFAULT
Default value of BATCH_ROWS_THRESHOLD_NAME
See Also:
@@ -1040,7 +1040,7 @@ implements
RESERVOIR_ENABLED_KEY
-protected static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String RESERVOIR_ENABLED_KEY
+protected static finalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String RESERVOIR_ENABLED_KEY
See Also:
Constant
Field Values
@@ -1053,7 +1053,7 @@ implements
requestCount
-finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true;
title="class or interface in java.util.concurrent.atomic">LongAdder requestCount
+finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true;
title="class or interface in java.util.concurrent.atomic">LongAdder requestCount
@@ -1062,7 +1062,7 @@ implements
rpcGetRequestCount
-finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true;
title="class or interface in java.util.concurrent.atomic">LongAdder rpcGetRequestCount
+finalhttps://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/LongAdder.html?is-external=true;
title="class or interface in
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/81cde4ce/devapidocs/org/apache/hadoop/hbase/master/ServerManager.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/master/ServerManager.html
b/devapidocs/org/apache/hadoop/hbase/master/ServerManager.html
index dbe6db5..0b18d24 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/ServerManager.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/ServerManager.html
@@ -353,7 +353,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
-void
+boolean
expireServer(ServerNameserverName)
@@ -1126,7 +1126,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
expireServer
-publicvoidexpireServer(ServerNameserverName)
+publicbooleanexpireServer(ServerNameserverName)
@@ -1135,7 +1135,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
moveFromOnlineToDeadServers
-publicvoidmoveFromOnlineToDeadServers(ServerNamesn)
+publicvoidmoveFromOnlineToDeadServers(ServerNamesn)
@@ -1144,7 +1144,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
processDeadServer
-publicvoidprocessDeadServer(ServerNameserverName,
+publicvoidprocessDeadServer(ServerNameserverName,
booleanshouldSplitWal)
@@ -1154,7 +1154,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
processQueuedDeadServers
-voidprocessQueuedDeadServers()
+voidprocessQueuedDeadServers()
Process the servers which died during master's
initialization. It will be
called after HMaster#assignMeta and AssignmentManager#joinCluster.
@@ -1165,7 +1165,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
removeServerFromDrainList
-publicbooleanremoveServerFromDrainList(ServerNamesn)
+publicbooleanremoveServerFromDrainList(ServerNamesn)
@@ -1174,7 +1174,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
addServerToDrainList
-publicbooleanaddServerToDrainList(ServerNamesn)
+publicbooleanaddServerToDrainList(ServerNamesn)
Add the server to the drain list.
Parameters:
@@ -1190,7 +1190,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
newRpcController
-privateHBaseRpcControllernewRpcController()
+privateHBaseRpcControllernewRpcController()
@@ -1199,7 +1199,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
sendRegionWarmup
-publicvoidsendRegionWarmup(ServerNameserver,
+publicvoidsendRegionWarmup(ServerNameserver,
RegionInforegion)
Sends a WARMUP RPC to the specified server to warmup the
specified region.
@@ -1218,7 +1218,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
closeRegionSilentlyAndWait
-public staticvoidcloseRegionSilentlyAndWait(ClusterConnectionconnection,
+public staticvoidcloseRegionSilentlyAndWait(ClusterConnectionconnection,
ServerNameserver,
RegionInforegion,
longtimeout)
@@ -1239,7 +1239,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
getRsAdmin
-publicorg.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService.BlockingInterfacegetRsAdmin(ServerNamesn)
+publicorg.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.AdminService.BlockingInterfacegetRsAdmin(ServerNamesn)
throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
Parameters:
@@ -1258,7 +1258,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
getMinToStart
-privateintgetMinToStart()
+privateintgetMinToStart()
Calculate min necessary to start. This is not an absolute.
It is just
a friction that will cause us hang around a bit longer waiting on
RegionServers to check-in.
@@ -1270,7 +1270,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
waitForRegionServers
-publicvoidwaitForRegionServers(MonitoredTaskstatus)
+publicvoidwaitForRegionServers(MonitoredTaskstatus)
throws https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true;
title="class or interface in java.lang">InterruptedException
Wait for the region servers to report in.
We will wait until one of this condition is met:
@@ -1293,7 +1293,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html
getStrForMax
-privatehttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">StringgetStrForMax(intmax)
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d347bde8/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html
b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html
index c0b3f3f..de426ec 100644
--- a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html
+++ b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/BlockCacheKey.html
@@ -168,27 +168,39 @@
void
+CombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
+ Cacheablebuf)
+
+
+void
BlockCache.cacheBlock(BlockCacheKeycacheKey,
Cacheablebuf)
Add block to cache (defaults to not in-memory).
-
+
void
LruBlockCache.cacheBlock(BlockCacheKeycacheKey,
Cacheablebuf)
Cache the block with the specified name and buffer.
-
+
void
-CombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
+MemcachedBlockCache.cacheBlock(BlockCacheKeycacheKey,
Cacheablebuf)
+
+void
+CombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
+ Cacheablebuf,
+ booleaninMemory)
+
void
-MemcachedBlockCache.cacheBlock(BlockCacheKeycacheKey,
- Cacheablebuf)
+InclusiveCombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
+ Cacheablebuf,
+ booleaninMemory)
void
@@ -208,18 +220,6 @@
void
-CombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
- Cacheablebuf,
- booleaninMemory)
-
-
-void
-InclusiveCombinedBlockCache.cacheBlock(BlockCacheKeycacheKey,
- Cacheablebuf,
- booleaninMemory)
-
-
-void
MemcachedBlockCache.cacheBlock(BlockCacheKeycacheKey,
Cacheablebuf,
booleaninMemory)
@@ -232,21 +232,21 @@
boolean
-BlockCache.evictBlock(BlockCacheKeycacheKey)
-Evict block from cache.
-
+CombinedBlockCache.evictBlock(BlockCacheKeycacheKey)
boolean
-LruBlockCache.evictBlock(BlockCacheKeycacheKey)
+InclusiveCombinedBlockCache.evictBlock(BlockCacheKeycacheKey)
boolean
-CombinedBlockCache.evictBlock(BlockCacheKeycacheKey)
+BlockCache.evictBlock(BlockCacheKeycacheKey)
+Evict block from cache.
+
boolean
-InclusiveCombinedBlockCache.evictBlock(BlockCacheKeycacheKey)
+LruBlockCache.evictBlock(BlockCacheKeycacheKey)
boolean
@@ -254,35 +254,35 @@
Cacheable
-BlockCache.getBlock(BlockCacheKeycacheKey,
+CombinedBlockCache.getBlock(BlockCacheKeycacheKey,
booleancaching,
booleanrepeat,
-booleanupdateCacheMetrics)
-Fetch block from cache.
-
+booleanupdateCacheMetrics)
Cacheable
-LruBlockCache.getBlock(BlockCacheKeycacheKey,
+InclusiveCombinedBlockCache.getBlock(BlockCacheKeycacheKey,
booleancaching,
booleanrepeat,
-booleanupdateCacheMetrics)
-Get the buffer of the block with the specified name.
-
+booleanupdateCacheMetrics)
Cacheable
-CombinedBlockCache.getBlock(BlockCacheKeycacheKey,
+BlockCache.getBlock(BlockCacheKeycacheKey,
booleancaching,
booleanrepeat,
-booleanupdateCacheMetrics)
+booleanupdateCacheMetrics)
+Fetch block from cache.
+
Cacheable
-InclusiveCombinedBlockCache.getBlock(BlockCacheKeycacheKey,
+LruBlockCache.getBlock(BlockCacheKeycacheKey,
booleancaching,
booleanrepeat,
-booleanupdateCacheMetrics)
+booleanupdateCacheMetrics)
+Get the buffer of the block with the specified name.
+
Cacheable
@@ -308,6 +308,11 @@
CombinedBlockCache.getRefCount(BlockCacheKeycacheKey)
+void
+CombinedBlockCache.returnBlock(BlockCacheKeycacheKey,
+ Cacheableblock)
+
+
default void
BlockCache.returnBlock(BlockCacheKeycacheKey,
Cacheableblock)
@@ -315,11 +320,6 @@
is over.
-
-void
-CombinedBlockCache.returnBlock(BlockCacheKeycacheKey,
- Cacheableblock)
-
@@ -497,13 +497,13 @@
void
-CachedEntryQueue.add(https://docs.oracle.com/javase/8/docs/api/java/util/Map.Entry.html?is-external=true;
title="class or interface in java.util">Map.EntryBlockCacheKey,BucketCache.BucketEntryentry)
-Attempt to add the specified entry to this queue.
-
+BucketCache.BucketEntryGroup.add(https://docs.oracle.com/javase/8/docs/api/java/util/Map.Entry.html?is-external=true;
title="class or interface in java.util">Map.EntryBlockCacheKey,BucketCache.BucketEntryblock)
void
-BucketCache.BucketEntryGroup.add(https://docs.oracle.com/javase/8/docs/api/java/util/Map.Entry.html?is-external=true;
title="class or interface in java.util">Map.EntryBlockCacheKey,BucketCache.BucketEntryblock)
+CachedEntryQueue.add(https://docs.oracle.com/javase/8/docs/api/java/util/Map.Entry.html?is-external=true;
title="class or interface in java.util">Map.EntryBlockCacheKey,BucketCache.BucketEntryentry)
+Attempt to add the specified entry to this queue.
+
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6b94a2f2/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterFileSystem.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterFileSystem.html
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterFileSystem.html
index c8098b1..019ceeb 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterFileSystem.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterFileSystem.html
@@ -216,245 +216,254 @@
208 /**
209 * @return HBase root log dir.
210 */
-211 public Path getWALRootDir() { return
this.walRootDir; }
-212
-213 /**
-214 * @return HBase temp dir.
-215 */
-216 public Path getTempDir() {
-217return this.tempdir;
-218 }
-219
-220 /**
-221 * @return The unique identifier
generated for this cluster
-222 */
-223 public ClusterId getClusterId() {
-224return clusterId;
-225 }
-226
-227 /**
-228 * Get the rootdir. Make sure its
wholesome and exists before returning.
-229 * @param rd
-230 * @param c
-231 * @param fs
-232 * @return hbase.rootdir (after checks
for existence and bootstrapping if
-233 * needed populating the directory with
necessary bootup files).
-234 * @throws IOException
-235 */
-236 private Path checkRootDir(final Path
rd, final Configuration c, final FileSystem fs)
-237 throws IOException {
-238// If FS is in safe mode wait till
out of it.
-239FSUtils.waitOnSafeMode(c,
c.getInt(HConstants.THREAD_WAKE_FREQUENCY, 10 * 1000));
-240
-241// Filesystem is good. Go ahead and
check for hbase.rootdir.
-242try {
-243 if (!fs.exists(rd)) {
-244fs.mkdirs(rd);
-245// DFS leaves safe mode with 0
DNs when there are 0 blocks.
-246// We used to handle this by
checking the current DN count and waiting until
-247// it is nonzero. With security,
the check for datanode count doesn't work --
-248// it is a privileged op. So
instead we adopt the strategy of the jobtracker
-249// and simply retry file creation
during bootstrap indefinitely. As soon as
-250// there is one datanode it will
succeed. Permission problems should have
-251// already been caught by mkdirs
above.
-252FSUtils.setVersion(fs, rd,
c.getInt(HConstants.THREAD_WAKE_FREQUENCY,
-253 10 * 1000),
c.getInt(HConstants.VERSION_FILE_WRITE_ATTEMPTS,
-254
HConstants.DEFAULT_VERSION_FILE_WRITE_ATTEMPTS));
-255 } else {
-256if (!fs.isDirectory(rd)) {
-257 throw new
IllegalArgumentException(rd.toString() + " is not a directory");
-258}
-259// as above
-260FSUtils.checkVersion(fs, rd,
true, c.getInt(HConstants.THREAD_WAKE_FREQUENCY,
-261 10 * 1000),
c.getInt(HConstants.VERSION_FILE_WRITE_ATTEMPTS,
-262
HConstants.DEFAULT_VERSION_FILE_WRITE_ATTEMPTS));
-263 }
-264} catch (DeserializationException de)
{
-265 LOG.error(HBaseMarkers.FATAL,
"Please fix invalid configuration for "
-266+ HConstants.HBASE_DIR, de);
-267 IOException ioe = new
IOException();
-268 ioe.initCause(de);
-269 throw ioe;
-270} catch (IllegalArgumentException
iae) {
-271 LOG.error(HBaseMarkers.FATAL,
"Please fix invalid configuration for "
-272+ HConstants.HBASE_DIR + " " +
rd.toString(), iae);
-273 throw iae;
-274}
-275// Make sure cluster ID exists
-276if (!FSUtils.checkClusterIdExists(fs,
rd, c.getInt(
-277HConstants.THREAD_WAKE_FREQUENCY,
10 * 1000))) {
-278 FSUtils.setClusterId(fs, rd, new
ClusterId(), c.getInt(HConstants.THREAD_WAKE_FREQUENCY, 10 * 1000));
-279}
-280clusterId = FSUtils.getClusterId(fs,
rd);
-281
-282// Make sure the meta region
directory exists!
-283if (!FSUtils.metaRegionExists(fs,
rd)) {
-284 bootstrap(rd, c);
-285}
-286
-287// Create tableinfo-s for hbase:meta
if not already there.
-288// assume, created table descriptor
is for enabling table
-289// meta table is a system table, so
descriptors are predefined,
-290// we should get them from
registry.
-291FSTableDescriptors fsd = new
FSTableDescriptors(c, fs, rd);
-292
fsd.createTableDescriptor(fsd.get(TableName.META_TABLE_NAME));
-293
-294return rd;
-295 }
-296
-297 /**
-298 * Make sure the hbase temp directory
exists and is empty.
-299 * NOTE that this method is only
executed once just after the master becomes the active one.
-300 */
-301 private void checkTempDir(final Path
tmpdir, final Configuration c, final FileSystem fs)
-302 throws IOException {
-303// If the temp directory exists,
clear the content (left over, from the previous run)
-304if (fs.exists(tmpdir)) {
-305 // Archive table in temp, maybe
left over from failed deletion,
-306 // if not the cleaner will take
care of them.
-307 for
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/eb05e3e3/apidocs/org/apache/hadoop/hbase/client/Scan.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/client/Scan.html
b/apidocs/org/apache/hadoop/hbase/client/Scan.html
index 806ad85..1ca681e 100644
--- a/apidocs/org/apache/hadoop/hbase/client/Scan.html
+++ b/apidocs/org/apache/hadoop/hbase/client/Scan.html
@@ -97,7 +97,7 @@ var activeTableTab = "activeTableTab";
-http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">java.lang.Object
+https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">java.lang.Object
org.apache.hadoop.hbase.client.Operation
@@ -215,19 +215,19 @@ extends
-static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
HBASE_CLIENT_SCANNER_ASYNC_PREFETCH
Parameter name for client scanner sync/async prefetch
toggle.
-static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
SCAN_ATTRIBUTES_METRICS_DATA
Deprecated.
-static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
SCAN_ATTRIBUTES_METRICS_ENABLE
Deprecated.
since 1.0.0. Use setScanMetricsEnabled(boolean)
@@ -235,7 +235,7 @@ extends
-static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
+static https://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
SCAN_ATTRIBUTES_TABLE_NAME
@@ -361,7 +361,7 @@ extends getFamilies()
-http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
title="class or interface in java.util">Mapbyte[],http://docs.oracle.com/javase/8/docs/api/java/util/NavigableSet.html?is-external=true;
title="class or interface in
java.util">NavigableSetbyte[]
+https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
title="class or interface in java.util">Mapbyte[],https://docs.oracle.com/javase/8/docs/api/java/util/NavigableSet.html?is-external=true;
title="class or interface in
java.util">NavigableSetbyte[]
getFamilyMap()
Getting the familyMap
@@ -371,7 +371,7 @@ extends getFilter()
-http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String,http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
+https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String,https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
getFingerprint()
Compile the table and column family (i.e.
@@ -442,7 +442,7 @@ extends includeStopRow()
-http://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
title="class or interface in java.lang">Boolean
+https://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
title="class or interface in java.lang">Boolean
isAsyncPrefetch()
@@ -493,11 +493,11 @@ extends
Scan
-setACL(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
title="class or interface in java.util">Maphttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in
java.lang">String,org.apache.hadoop.hbase.security.access.Permissionperms)
+setACL(https://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
title="class or interface in java.util">Maphttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in
java.lang">String,org.apache.hadoop.hbase.security.access.Permissionperms)
Scan
-setACL(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">Stringuser,
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/ec8bf761/devapidocs/src-html/org/apache/hadoop/hbase/filter/CompareFilter.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/filter/CompareFilter.html
b/devapidocs/src-html/org/apache/hadoop/hbase/filter/CompareFilter.html
index e5fdac5..ad7c82a 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/filter/CompareFilter.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/filter/CompareFilter.html
@@ -33,299 +33,303 @@
025import org.apache.hadoop.hbase.Cell;
026import
org.apache.hadoop.hbase.CompareOperator;
027import
org.apache.hadoop.hbase.PrivateCellUtil;
-028import
org.apache.yetus.audience.InterfaceAudience;
-029import
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-030import
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
-031import
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
-032import
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.CompareType;
-033import
org.apache.hadoop.hbase.util.Bytes;
-034
-035import
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
-036/**
-037 * This is a generic filter to be used to
filter by comparison. It takes an
-038 * operator (equal, greater, not equal,
etc) and a byte [] comparator.
-039 * p
-040 * To filter by row key, use {@link
RowFilter}.
+028import
org.apache.hadoop.hbase.util.Bytes;
+029import
org.apache.yetus.audience.InterfaceAudience;
+030
+031import
org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
+032
+033import
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
+034import
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
+035import
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
+036import
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.CompareType;
+037
+038/**
+039 * This is a generic filter to be used to
filter by comparison. It takes an
+040 * operator (equal, greater, not equal,
etc) and a byte [] comparator.
041 * p
-042 * To filter by column qualifier, use
{@link QualifierFilter}.
+042 * To filter by row key, use {@link
RowFilter}.
043 * p
-044 * To filter by value, use {@link
SingleColumnValueFilter}.
+044 * To filter by column family, use {@link
FamilyFilter}.
045 * p
-046 * These filters can be wrapped with
{@link SkipFilter} and {@link WhileMatchFilter}
-047 * to add more control.
-048 * p
-049 * Multiple filters can be combined using
{@link FilterList}.
-050 */
-051@InterfaceAudience.Public
-052public abstract class CompareFilter
extends FilterBase {
-053 /**
-054 * Comparison operators. For filters
only!
-055 * Use {@link CompareOperator}
otherwise.
-056 * It (intentionally) has at least the
below enums with same names.
-057 * @deprecated since 2.0.0. Will be
removed in 3.0.0. Use {@link CompareOperator} instead.
-058 */
-059 @Deprecated
-060 @InterfaceAudience.Public
-061 public enum CompareOp {
-062/** less than */
-063LESS,
-064/** less than or equal to */
-065LESS_OR_EQUAL,
-066/** equals */
-067EQUAL,
-068/** not equal */
-069NOT_EQUAL,
-070/** greater than or equal to */
-071GREATER_OR_EQUAL,
-072/** greater than */
-073GREATER,
-074/** no operation */
-075NO_OP,
-076 }
-077
-078 protected CompareOperator op;
-079 protected ByteArrayComparable
comparator;
-080
-081 /**
-082 * Constructor.
-083 * @param compareOp the compare op for
row matching
-084 * @param comparator the comparator for
row matching
-085 * @deprecated Since 2.0.0. Will be
removed in 3.0.0. Use other constructor.
-086 */
-087 @Deprecated
-088 public CompareFilter(final CompareOp
compareOp,
-089 final ByteArrayComparable
comparator) {
-090
this(CompareOperator.valueOf(compareOp.name()), comparator);
-091 }
-092
-093 /**
-094 * Constructor.
-095 * @param op the compare op for row
matching
-096 * @param comparator the comparator for
row matching
-097 */
-098 public CompareFilter(final
CompareOperator op,
-099 final
ByteArrayComparable comparator) {
-100this.op = op;
-101this.comparator = comparator;
-102 }
-103
-104 /**
-105 * @return operator
-106 * @deprecated since 2.0.0. Will be
removed in 3.0.0. Use {@link #getCompareOperator()} instead.
-107 */
-108 @Deprecated
-109 public CompareOp getOperator() {
-110return
CompareOp.valueOf(op.name());
-111 }
-112
-113 public CompareOperator
getCompareOperator() {
-114return op;
+046 * To filter by column qualifier, use
{@link QualifierFilter}.
+047 * p
+048 * To filter by value, use {@link
ValueFilter}.
+049 * p
+050 * These filters can be wrapped with
{@link SkipFilter} and {@link WhileMatchFilter}
+051 * to add more control.
+052 * p
+053 * Multiple filters can be combined using
{@link FilterList}.
+054 */
+055@InterfaceAudience.Public
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/991224b9/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
index 802b925..a3e80ab 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.RegionScannerImpl.html
@@ -73,229 +73,229 @@
065import
java.util.concurrent.TimeoutException;
066import
java.util.concurrent.atomic.AtomicBoolean;
067import
java.util.concurrent.atomic.AtomicInteger;
-068import
java.util.concurrent.atomic.AtomicLong;
-069import
java.util.concurrent.atomic.LongAdder;
-070import java.util.concurrent.locks.Lock;
-071import
java.util.concurrent.locks.ReadWriteLock;
-072import
java.util.concurrent.locks.ReentrantReadWriteLock;
-073import java.util.function.Function;
-074import
org.apache.hadoop.conf.Configuration;
-075import org.apache.hadoop.fs.FileStatus;
-076import org.apache.hadoop.fs.FileSystem;
-077import
org.apache.hadoop.fs.LocatedFileStatus;
-078import org.apache.hadoop.fs.Path;
-079import org.apache.hadoop.hbase.Cell;
-080import
org.apache.hadoop.hbase.CellBuilderType;
-081import
org.apache.hadoop.hbase.CellComparator;
-082import
org.apache.hadoop.hbase.CellComparatorImpl;
-083import
org.apache.hadoop.hbase.CellScanner;
-084import
org.apache.hadoop.hbase.CellUtil;
-085import
org.apache.hadoop.hbase.CompareOperator;
-086import
org.apache.hadoop.hbase.CompoundConfiguration;
-087import
org.apache.hadoop.hbase.DoNotRetryIOException;
-088import
org.apache.hadoop.hbase.DroppedSnapshotException;
-089import
org.apache.hadoop.hbase.ExtendedCellBuilderFactory;
-090import
org.apache.hadoop.hbase.HConstants;
-091import
org.apache.hadoop.hbase.HConstants.OperationStatusCode;
-092import
org.apache.hadoop.hbase.HDFSBlocksDistribution;
-093import
org.apache.hadoop.hbase.HRegionInfo;
-094import
org.apache.hadoop.hbase.KeyValue;
-095import
org.apache.hadoop.hbase.KeyValueUtil;
-096import
org.apache.hadoop.hbase.NamespaceDescriptor;
-097import
org.apache.hadoop.hbase.NotServingRegionException;
-098import
org.apache.hadoop.hbase.PrivateCellUtil;
-099import
org.apache.hadoop.hbase.RegionTooBusyException;
-100import
org.apache.hadoop.hbase.TableName;
-101import org.apache.hadoop.hbase.Tag;
-102import org.apache.hadoop.hbase.TagUtil;
-103import
org.apache.hadoop.hbase.UnknownScannerException;
-104import
org.apache.hadoop.hbase.client.Append;
-105import
org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
-106import
org.apache.hadoop.hbase.client.CompactionState;
-107import
org.apache.hadoop.hbase.client.Delete;
-108import
org.apache.hadoop.hbase.client.Durability;
-109import
org.apache.hadoop.hbase.client.Get;
-110import
org.apache.hadoop.hbase.client.Increment;
-111import
org.apache.hadoop.hbase.client.IsolationLevel;
-112import
org.apache.hadoop.hbase.client.Mutation;
-113import
org.apache.hadoop.hbase.client.PackagePrivateFieldAccessor;
-114import
org.apache.hadoop.hbase.client.Put;
-115import
org.apache.hadoop.hbase.client.RegionInfo;
-116import
org.apache.hadoop.hbase.client.RegionReplicaUtil;
-117import
org.apache.hadoop.hbase.client.Result;
-118import
org.apache.hadoop.hbase.client.RowMutations;
-119import
org.apache.hadoop.hbase.client.Scan;
-120import
org.apache.hadoop.hbase.client.TableDescriptor;
-121import
org.apache.hadoop.hbase.client.TableDescriptorBuilder;
-122import
org.apache.hadoop.hbase.conf.ConfigurationManager;
-123import
org.apache.hadoop.hbase.conf.PropagatingConfigurationObserver;
-124import
org.apache.hadoop.hbase.coprocessor.RegionObserver.MutationType;
-125import
org.apache.hadoop.hbase.errorhandling.ForeignExceptionSnare;
-126import
org.apache.hadoop.hbase.exceptions.FailedSanityCheckException;
-127import
org.apache.hadoop.hbase.exceptions.TimeoutIOException;
-128import
org.apache.hadoop.hbase.exceptions.UnknownProtocolException;
-129import
org.apache.hadoop.hbase.filter.ByteArrayComparable;
-130import
org.apache.hadoop.hbase.filter.FilterWrapper;
-131import
org.apache.hadoop.hbase.filter.IncompatibleFilterException;
-132import
org.apache.hadoop.hbase.io.HFileLink;
-133import
org.apache.hadoop.hbase.io.HeapSize;
-134import
org.apache.hadoop.hbase.io.TimeRange;
-135import
org.apache.hadoop.hbase.io.hfile.HFile;
-136import
org.apache.hadoop.hbase.ipc.CallerDisconnectedException;
-137import
org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils;
-138import
org.apache.hadoop.hbase.ipc.RpcCall;
-139import
org.apache.hadoop.hbase.ipc.RpcServer;
-140import
org.apache.hadoop.hbase.monitoring.MonitoredTask;
-141import
org.apache.hadoop.hbase.monitoring.TaskMonitor;
-142import
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/193b4259/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.html
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.html
index bd13b53..802b925 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.html
@@ -900,7600 +900,7598 @@
892if
(this.getRegionInfo().getReplicaId() == RegionInfo.DEFAULT_REPLICA_ID) {
893 status.setStatus("Writing region
info on filesystem");
894 fs.checkRegionInfoOnFilesystem();
-895} else {
-896 if (LOG.isDebugEnabled()) {
-897LOG.debug("Skipping creation of
.regioninfo file for " + this.getRegionInfo());
-898 }
-899}
-900
-901// Initialize all the HStores
-902status.setStatus("Initializing all
the Stores");
-903long maxSeqId =
initializeStores(reporter, status);
-904this.mvcc.advanceTo(maxSeqId);
-905if
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
-906 CollectionHStore stores =
this.stores.values();
-907 try {
-908// update the stores that we are
replaying
-909
stores.forEach(HStore::startReplayingFromWAL);
-910// Recover any edits if
available.
-911maxSeqId = Math.max(maxSeqId,
-912
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter,
status));
-913// Make sure mvcc is up to max.
-914this.mvcc.advanceTo(maxSeqId);
-915 } finally {
-916// update the stores that we are
done replaying
-917
stores.forEach(HStore::stopReplayingFromWAL);
-918 }
-919}
-920this.lastReplayedOpenRegionSeqId =
maxSeqId;
+895}
+896
+897// Initialize all the HStores
+898status.setStatus("Initializing all
the Stores");
+899long maxSeqId =
initializeStores(reporter, status);
+900this.mvcc.advanceTo(maxSeqId);
+901if
(ServerRegionReplicaUtil.shouldReplayRecoveredEdits(this)) {
+902 CollectionHStore stores =
this.stores.values();
+903 try {
+904// update the stores that we are
replaying
+905
stores.forEach(HStore::startReplayingFromWAL);
+906// Recover any edits if
available.
+907maxSeqId = Math.max(maxSeqId,
+908
replayRecoveredEditsIfAny(this.fs.getRegionDir(), maxSeqIdInStores, reporter,
status));
+909// Make sure mvcc is up to max.
+910this.mvcc.advanceTo(maxSeqId);
+911 } finally {
+912// update the stores that we are
done replaying
+913
stores.forEach(HStore::stopReplayingFromWAL);
+914 }
+915}
+916this.lastReplayedOpenRegionSeqId =
maxSeqId;
+917
+918
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
+919this.writestate.flushRequested =
false;
+920this.writestate.compacting.set(0);
921
-922
this.writestate.setReadOnly(ServerRegionReplicaUtil.isReadOnly(this));
-923this.writestate.flushRequested =
false;
-924this.writestate.compacting.set(0);
-925
-926if (this.writestate.writesEnabled)
{
-927 // Remove temporary data left over
from old regions
-928 status.setStatus("Cleaning up
temporary data from old regions");
-929 fs.cleanupTempDir();
-930}
-931
-932if (this.writestate.writesEnabled)
{
-933 status.setStatus("Cleaning up
detritus from prior splits");
-934 // Get rid of any splits or merges
that were lost in-progress. Clean out
-935 // these directories here on open.
We may be opening a region that was
-936 // being split but we crashed in
the middle of it all.
-937 fs.cleanupAnySplitDetritus();
-938 fs.cleanupMergesDir();
-939}
-940
-941// Initialize split policy
-942this.splitPolicy =
RegionSplitPolicy.create(this, conf);
-943
-944// Initialize flush policy
-945this.flushPolicy =
FlushPolicyFactory.create(this, conf);
-946
-947long lastFlushTime =
EnvironmentEdgeManager.currentTime();
-948for (HStore store: stores.values())
{
-949
this.lastStoreFlushTimeMap.put(store, lastFlushTime);
-950}
-951
-952// Use maximum of log sequenceid or
that which was found in stores
-953// (particularly if no recovered
edits, seqid will be -1).
-954long nextSeqid = maxSeqId;
-955if (this.writestate.writesEnabled)
{
-956 nextSeqid =
WALSplitter.writeRegionSequenceIdFile(this.fs.getFileSystem(),
-957 this.fs.getRegionDir(),
nextSeqid, 1);
-958} else {
-959 nextSeqid++;
-960}
-961
-962LOG.info("Onlined " +
this.getRegionInfo().getShortNameToLog() +
-963 "; next sequenceid=" +
nextSeqid);
+922if (this.writestate.writesEnabled)
{
+923 // Remove temporary data left over
from old regions
+924 status.setStatus("Cleaning up
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/94208cfe/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html
b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html
index 554f69f..e901330 100644
--- a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html
+++ b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/Procedure.html
@@ -132,13 +132,13 @@
http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListProcedure?
-MasterServices.getProcedures()
-Get procedures
-
+HMaster.getProcedures()
http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListProcedure?
-HMaster.getProcedures()
+MasterServices.getProcedures()
+Get procedures
+
@@ -531,20 +531,20 @@
Procedureproc)
-boolean
-PeerQueue.requireExclusiveLock(Procedure?proc)
-
-
(package private) abstract boolean
Queue.requireExclusiveLock(Procedure?proc)
+
+boolean
+ServerQueue.requireExclusiveLock(Procedure?proc)
+
boolean
-TableQueue.requireExclusiveLock(Procedure?proc)
+PeerQueue.requireExclusiveLock(Procedure?proc)
boolean
-ServerQueue.requireExclusiveLock(Procedure?proc)
+TableQueue.requireExclusiveLock(Procedure?proc)
static byte[]
@@ -920,44 +920,44 @@
-protected Procedure
-SimpleProcedureScheduler.dequeue()
-
-
protected abstract Procedure
AbstractProcedureScheduler.dequeue()
Fetch one Procedure from the queue
NOTE: this method is called with the sched lock held.
-
-protected Procedure[]
-SequentialProcedure.doExecute(TEnvironmentenv)
-
+protected Procedure
+SimpleProcedureScheduler.dequeue()
+
+
protected ProcedureTEnvironment[]
Procedure.doExecute(TEnvironmentenv)
Internal method called by the ProcedureExecutor that starts
the user-level code execute().
-
-protected ProcedureTEnvironment[]
-ProcedureExecutor.FailedProcedure.execute(TEnvironmentenv)
-
protected Procedure[]
+SequentialProcedure.doExecute(TEnvironmentenv)
+
+
+protected Procedure[]
StateMachineProcedure.execute(TEnvironmentenv)
+
+protected ProcedureTEnvironment[]
+ProcedureInMemoryChore.execute(TEnvironmentenv)
+
+protected ProcedureTEnvironment[]
+ProcedureExecutor.FailedProcedure.execute(TEnvironmentenv)
+
+
protected abstract ProcedureTEnvironment[]
Procedure.execute(TEnvironmentenv)
The main code of the procedure.
-
-protected ProcedureTEnvironment[]
-ProcedureInMemoryChore.execute(TEnvironmentenv)
-
Procedure?
LockedResource.getExclusiveLockOwnerProcedure()
@@ -1115,14 +1115,14 @@
void
-SimpleProcedureScheduler.completionCleanup(Procedureproc)
-
-
-void
ProcedureScheduler.completionCleanup(Procedureproc)
The procedure in execution completed.
+
+void
+SimpleProcedureScheduler.completionCleanup(Procedureproc)
+
static
org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure
ProcedureUtil.convertToProtoProcedure(Procedureproc)
@@ -1135,17 +1135,17 @@
Procedureprocedure)
-protected void
-SimpleProcedureScheduler.enqueue(Procedureprocedure,
- booleanaddFront)
-
-
protected abstract void
AbstractProcedureScheduler.enqueue(Procedureprocedure,
booleanaddFront)
Add the procedure to the queue.
+
+protected void
+SimpleProcedureScheduler.enqueue(Procedureprocedure,
+ booleanaddFront)
+
private void
ProcedureExecutor.execCompletionCleanup(Procedureproc)
@@ -1327,14 +1327,14 @@
void
-SimpleProcedureScheduler.yield(Procedureproc)
-
-
-void
ProcedureScheduler.yield(Procedureproc)
The procedure can't run at the moment.
+
+void
+SimpleProcedureScheduler.yield(Procedureproc)
+
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/94208cfe/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureEvent.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureEvent.html
b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureEvent.html
index a9a3870..5f35947 100644
---
a/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureEvent.html
+++
b/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureEvent.html
@@ -141,11 +141,11 @@
ProcedureEvent?
-MasterServices.getInitializedEvent()
+HMaster.getInitializedEvent()
ProcedureEvent?
-HMaster.getInitializedEvent()
+MasterServices.getInitializedEvent()
ProcedureEvent?
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/94208cfe/devapidocs/org/apache/hadoop/hbase/procedure2/class-use/ProcedureExecutor.html
--
diff --git
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cd17dc5/devapidocs/org/apache/hadoop/hbase/ipc/class-use/PriorityFunction.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/PriorityFunction.html
b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/PriorityFunction.html
index 7244ce2..5f7ce59 100644
--- a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/PriorityFunction.html
+++ b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/PriorityFunction.html
@@ -114,15 +114,15 @@
private PriorityFunction
-SimpleRpcScheduler.priority
+RpcExecutor.priority
private PriorityFunction
-RpcExecutor.priority
+RpcExecutor.CallPriorityComparator.priority
private PriorityFunction
-RpcExecutor.CallPriorityComparator.priority
+SimpleRpcScheduler.priority
@@ -319,7 +319,7 @@
RpcScheduler
-RpcSchedulerFactory.create(org.apache.hadoop.conf.Configurationconf,
+FifoRpcSchedulerFactory.create(org.apache.hadoop.conf.Configurationconf,
PriorityFunctionpriority)
Deprecated.
@@ -333,18 +333,16 @@
RpcScheduler
-FifoRpcSchedulerFactory.create(org.apache.hadoop.conf.Configurationconf,
+RpcSchedulerFactory.create(org.apache.hadoop.conf.Configurationconf,
PriorityFunctionpriority)
Deprecated.
RpcScheduler
-RpcSchedulerFactory.create(org.apache.hadoop.conf.Configurationconf,
+FifoRpcSchedulerFactory.create(org.apache.hadoop.conf.Configurationconf,
PriorityFunctionpriority,
- Abortableserver)
-Constructs a RpcScheduler.
-
+ Abortableserver)
RpcScheduler
@@ -354,9 +352,11 @@
RpcScheduler
-FifoRpcSchedulerFactory.create(org.apache.hadoop.conf.Configurationconf,
+RpcSchedulerFactory.create(org.apache.hadoop.conf.Configurationconf,
PriorityFunctionpriority,
- Abortableserver)
+ Abortableserver)
+Constructs a RpcScheduler.
+
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cd17dc5/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcCallback.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcCallback.html
b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcCallback.html
index 4a25f5c..6d59fb7 100644
--- a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcCallback.html
+++ b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcCallback.html
@@ -123,14 +123,14 @@
void
-ServerCall.setCallBack(RpcCallbackcallback)
-
-
-void
RpcCallContext.setCallBack(RpcCallbackcallback)
Sets a callback which has to be executed at the end of this
RPC call.
+
+void
+ServerCall.setCallBack(RpcCallbackcallback)
+
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cd17dc5/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcControllerFactory.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcControllerFactory.html
b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcControllerFactory.html
index fab4d7a..baa4e5e 100644
--- a/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcControllerFactory.html
+++ b/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcControllerFactory.html
@@ -131,32 +131,24 @@
-private RpcControllerFactory
-ConnectionImplementation.rpcControllerFactory
-
-
-protected RpcControllerFactory
-ClientScanner.rpcControllerFactory
-
-
protected RpcControllerFactory
RegionAdminServiceCallable.rpcControllerFactory
-(package private) RpcControllerFactory
-AsyncConnectionImpl.rpcControllerFactory
+private RpcControllerFactory
+ConnectionImplementation.rpcControllerFactory
-private RpcControllerFactory
-HTable.rpcControllerFactory
+(package private) RpcControllerFactory
+AsyncConnectionImpl.rpcControllerFactory
private RpcControllerFactory
-HBaseAdmin.rpcControllerFactory
+HTable.rpcControllerFactory
private RpcControllerFactory
-SecureBulkLoadClient.rpcControllerFactory
+RpcRetryingCallerWithReadReplicas.rpcControllerFactory
protected RpcControllerFactory
@@ -164,7 +156,15 @@
private RpcControllerFactory
-RpcRetryingCallerWithReadReplicas.rpcControllerFactory
+HBaseAdmin.rpcControllerFactory
+
+
+private RpcControllerFactory
+SecureBulkLoadClient.rpcControllerFactory
+
+
+protected RpcControllerFactory
+ClientScanner.rpcControllerFactory
(package private) RpcControllerFactory
@@ -181,11 +181,11 @@
RpcControllerFactory
-ConnectionImplementation.getRpcControllerFactory()
+ClusterConnection.getRpcControllerFactory()
RpcControllerFactory
-ClusterConnection.getRpcControllerFactory()
+ConnectionImplementation.getRpcControllerFactory()
private RpcControllerFactory
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0cd17dc5/devapidocs/org/apache/hadoop/hbase/ipc/class-use/RpcExecutor.Handler.html
--
diff --git
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
index 3628d68..bd2f966 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
@@ -152,27 +152,27 @@ the order they are declared.
PeerProcedureInterface.PeerOperationType
-RefreshPeerProcedure.getPeerOperationType()
+DisablePeerProcedure.getPeerOperationType()
PeerProcedureInterface.PeerOperationType
-DisablePeerProcedure.getPeerOperationType()
+RemovePeerProcedure.getPeerOperationType()
PeerProcedureInterface.PeerOperationType
-UpdatePeerConfigProcedure.getPeerOperationType()
+EnablePeerProcedure.getPeerOperationType()
PeerProcedureInterface.PeerOperationType
-AddPeerProcedure.getPeerOperationType()
+RefreshPeerProcedure.getPeerOperationType()
PeerProcedureInterface.PeerOperationType
-EnablePeerProcedure.getPeerOperationType()
+AddPeerProcedure.getPeerOperationType()
PeerProcedureInterface.PeerOperationType
-RemovePeerProcedure.getPeerOperationType()
+UpdatePeerConfigProcedure.getPeerOperationType()
private static PeerProcedureInterface.PeerOperationType
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
index 269bc46..f7a6279 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
@@ -125,11 +125,11 @@
private ProcedurePrepareLatch
-RecoverMetaProcedure.syncLatch
+AbstractStateMachineTableProcedure.syncLatch
private ProcedurePrepareLatch
-AbstractStateMachineTableProcedure.syncLatch
+RecoverMetaProcedure.syncLatch
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
index 5e8085c..8b6ceb7 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
@@ -104,14 +104,14 @@
ServerProcedureInterface.ServerOperationType
-ServerCrashProcedure.getServerOperationType()
-
-
-ServerProcedureInterface.ServerOperationType
ServerProcedureInterface.getServerOperationType()
Given an operation type we can take decisions about what to
do with pending operations.
+
+ServerProcedureInterface.ServerOperationType
+ServerCrashProcedure.getServerOperationType()
+
static ServerProcedureInterface.ServerOperationType
ServerProcedureInterface.ServerOperationType.valueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">Stringname)
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/828486ae/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
index 046295e..e736f37 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
@@ -112,19 +112,19 @@
TableProcedureInterface.TableOperationType
-MoveRegionProcedure.getTableOperationType()
+UnassignProcedure.getTableOperationType()
TableProcedureInterface.TableOperationType
-GCMergedRegionsProcedure.getTableOperationType()
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFile.Writer.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFile.Writer.html
b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFile.Writer.html
index abcb738..c7d05d1 100644
--- a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFile.Writer.html
+++ b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFile.Writer.html
@@ -143,17 +143,17 @@
void
-NoOpDataBlockEncoder.saveMetadata(HFile.Writerwriter)
+HFileDataBlockEncoderImpl.saveMetadata(HFile.Writerwriter)
void
-HFileDataBlockEncoder.saveMetadata(HFile.Writerwriter)
-Save metadata in HFile which will be written to disk
-
+NoOpDataBlockEncoder.saveMetadata(HFile.Writerwriter)
void
-HFileDataBlockEncoderImpl.saveMetadata(HFile.Writerwriter)
+HFileDataBlockEncoder.saveMetadata(HFile.Writerwriter)
+Save metadata in HFile which will be written to disk
+
@@ -203,18 +203,18 @@
-abstract void
-BloomContext.addLastBloomKey(HFile.Writerwriter)
-Adds the last bloom key to the HFile Writer as part of
StorefileWriter close.
-
+void
+RowColBloomContext.addLastBloomKey(HFile.Writerwriter)
void
RowBloomContext.addLastBloomKey(HFile.Writerwriter)
-void
-RowColBloomContext.addLastBloomKey(HFile.Writerwriter)
+abstract void
+BloomContext.addLastBloomKey(HFile.Writerwriter)
+Adds the last bloom key to the HFile Writer as part of
StorefileWriter close.
+
static BloomFilterWriter
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileBlock.Writer.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileBlock.Writer.html
b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileBlock.Writer.html
index 274bfad..479b9d3 100644
---
a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileBlock.Writer.html
+++
b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileBlock.Writer.html
@@ -106,15 +106,15 @@
-private HFileBlock.Writer
-HFileBlockIndex.BlockIndexWriter.blockWriter
-
-
protected HFileBlock.Writer
HFileWriterImpl.blockWriter
block writer
+
+private HFileBlock.Writer
+HFileBlockIndex.BlockIndexWriter.blockWriter
+
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f272b0e8/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileContext.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileContext.html
b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileContext.html
index b293c97..0c892c8 100644
--- a/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileContext.html
+++ b/devapidocs/org/apache/hadoop/hbase/io/hfile/class-use/HFileContext.html
@@ -136,15 +136,15 @@
HFileContext
-HFileBlockEncodingContext.getHFileContext()
+HFileBlockDecodingContext.getHFileContext()
HFileContext
-HFileBlockDecodingContext.getHFileContext()
+HFileBlockDefaultDecodingContext.getHFileContext()
HFileContext
-HFileBlockDefaultDecodingContext.getHFileContext()
+HFileBlockEncodingContext.getHFileContext()
HFileContext
@@ -224,24 +224,24 @@
private HFileContext
-HFile.WriterFactory.fileContext
-
-
-private HFileContext
HFileBlock.fileContext
Meta data that holds meta information on the
hfileblock.
-
+
private HFileContext
HFileBlock.Writer.fileContext
Meta data that holds information about the hfileblock
-
+
private HFileContext
HFileBlock.FSReaderImpl.fileContext
+
+private HFileContext
+HFile.WriterFactory.fileContext
+
private HFileContext
HFileReaderImpl.hfileContext
@@ -277,20 +277,20 @@
HFileContext
+HFileWriterImpl.getFileContext()
+
+
+HFileContext
HFile.Writer.getFileContext()
Return the file context for the HFile this writer belongs
to
-
+
HFileContext
HFile.Reader.getFileContext()
Return the file context of the HFile this reader belongs
to
-
-HFileContext
-HFileWriterImpl.getFileContext()
-
HFileContext
HFileReaderImpl.getFileContext()
@@ -323,35 +323,35 @@
HFileBlockDecodingContext
-NoOpDataBlockEncoder.newDataBlockDecodingContext(HFileContextmeta)
+HFileDataBlockEncoderImpl.newDataBlockDecodingContext(HFileContextfileContext)
HFileBlockDecodingContext
-HFileDataBlockEncoder.newDataBlockDecodingContext(HFileContextfileContext)
-create a encoder specific decoding context for
reading.
-
+NoOpDataBlockEncoder.newDataBlockDecodingContext(HFileContextmeta)
HFileBlockDecodingContext
-HFileDataBlockEncoderImpl.newDataBlockDecodingContext(HFileContextfileContext)
+HFileDataBlockEncoder.newDataBlockDecodingContext(HFileContextfileContext)
+create a encoder specific decoding context for
reading.
+
HFileBlockEncodingContext
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
index 3628d68..bd2f966 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/PeerProcedureInterface.PeerOperationType.html
@@ -152,27 +152,27 @@ the order they are declared.
PeerProcedureInterface.PeerOperationType
-RefreshPeerProcedure.getPeerOperationType()
+DisablePeerProcedure.getPeerOperationType()
PeerProcedureInterface.PeerOperationType
-DisablePeerProcedure.getPeerOperationType()
+RemovePeerProcedure.getPeerOperationType()
PeerProcedureInterface.PeerOperationType
-UpdatePeerConfigProcedure.getPeerOperationType()
+EnablePeerProcedure.getPeerOperationType()
PeerProcedureInterface.PeerOperationType
-AddPeerProcedure.getPeerOperationType()
+RefreshPeerProcedure.getPeerOperationType()
PeerProcedureInterface.PeerOperationType
-EnablePeerProcedure.getPeerOperationType()
+AddPeerProcedure.getPeerOperationType()
PeerProcedureInterface.PeerOperationType
-RemovePeerProcedure.getPeerOperationType()
+UpdatePeerConfigProcedure.getPeerOperationType()
private static PeerProcedureInterface.PeerOperationType
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
index 269bc46..f7a6279 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ProcedurePrepareLatch.html
@@ -125,11 +125,11 @@
private ProcedurePrepareLatch
-RecoverMetaProcedure.syncLatch
+AbstractStateMachineTableProcedure.syncLatch
private ProcedurePrepareLatch
-AbstractStateMachineTableProcedure.syncLatch
+RecoverMetaProcedure.syncLatch
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
index 5e8085c..8b6ceb7 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/ServerProcedureInterface.ServerOperationType.html
@@ -104,14 +104,14 @@
ServerProcedureInterface.ServerOperationType
-ServerCrashProcedure.getServerOperationType()
-
-
-ServerProcedureInterface.ServerOperationType
ServerProcedureInterface.getServerOperationType()
Given an operation type we can take decisions about what to
do with pending operations.
+
+ServerProcedureInterface.ServerOperationType
+ServerCrashProcedure.getServerOperationType()
+
static ServerProcedureInterface.ServerOperationType
ServerProcedureInterface.ServerOperationType.valueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">Stringname)
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c83a37c8/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
index 046295e..e736f37 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/procedure/class-use/TableProcedureInterface.TableOperationType.html
@@ -112,19 +112,19 @@
TableProcedureInterface.TableOperationType
-MoveRegionProcedure.getTableOperationType()
+UnassignProcedure.getTableOperationType()
TableProcedureInterface.TableOperationType
-GCMergedRegionsProcedure.getTableOperationType()
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0ab8335e/devapidocs/org/apache/hadoop/hbase/nio/class-use/ByteBuff.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/nio/class-use/ByteBuff.html
b/devapidocs/org/apache/hadoop/hbase/nio/class-use/ByteBuff.html
index 158a6aa..74a0687 100644
--- a/devapidocs/org/apache/hadoop/hbase/nio/class-use/ByteBuff.html
+++ b/devapidocs/org/apache/hadoop/hbase/nio/class-use/ByteBuff.html
@@ -161,23 +161,23 @@
Codec.Decoder
-KeyValueCodec.getDecoder(ByteBuffbuf)
+CellCodec.getDecoder(ByteBuffbuf)
Codec.Decoder
-CellCodecWithTags.getDecoder(ByteBuffbuf)
+Codec.getDecoder(ByteBuffbuf)
Codec.Decoder
-Codec.getDecoder(ByteBuffbuf)
+KeyValueCodec.getDecoder(ByteBuffbuf)
Codec.Decoder
-CellCodec.getDecoder(ByteBuffbuf)
+KeyValueCodecWithTags.getDecoder(ByteBuffbuf)
Codec.Decoder
-KeyValueCodecWithTags.getDecoder(ByteBuffbuf)
+CellCodecWithTags.getDecoder(ByteBuffbuf)
Codec.Decoder
@@ -259,20 +259,20 @@
-protected ByteBuff
-BufferedDataBlockEncoder.SeekerState.currentBuffer
+private ByteBuff
+RowIndexSeekerV1.currentBuffer
protected ByteBuff
-BufferedDataBlockEncoder.BufferedEncodedSeeker.currentBuffer
+RowIndexSeekerV1.SeekerState.currentBuffer
-private ByteBuff
-RowIndexSeekerV1.currentBuffer
+protected ByteBuff
+BufferedDataBlockEncoder.SeekerState.currentBuffer
protected ByteBuff
-RowIndexSeekerV1.SeekerState.currentBuffer
+BufferedDataBlockEncoder.BufferedEncodedSeeker.currentBuffer
private ByteBuff
@@ -295,23 +295,23 @@
Cell
-CopyKeyDataBlockEncoder.getFirstKeyCellInBlock(ByteBuffblock)
+RowIndexCodecV1.getFirstKeyCellInBlock(ByteBuffblock)
Cell
-PrefixKeyDeltaEncoder.getFirstKeyCellInBlock(ByteBuffblock)
+CopyKeyDataBlockEncoder.getFirstKeyCellInBlock(ByteBuffblock)
Cell
-FastDiffDeltaEncoder.getFirstKeyCellInBlock(ByteBuffblock)
+DiffKeyDeltaEncoder.getFirstKeyCellInBlock(ByteBuffblock)
Cell
-DiffKeyDeltaEncoder.getFirstKeyCellInBlock(ByteBuffblock)
+FastDiffDeltaEncoder.getFirstKeyCellInBlock(ByteBuffblock)
Cell
-RowIndexCodecV1.getFirstKeyCellInBlock(ByteBuffblock)
+PrefixKeyDeltaEncoder.getFirstKeyCellInBlock(ByteBuffblock)
void
@@ -338,11 +338,11 @@
void
-BufferedDataBlockEncoder.BufferedEncodedSeeker.setCurrentBuffer(ByteBuffbuffer)
+RowIndexSeekerV1.setCurrentBuffer(ByteBuffbuffer)
void
-RowIndexSeekerV1.setCurrentBuffer(ByteBuffbuffer)
+BufferedDataBlockEncoder.BufferedEncodedSeeker.setCurrentBuffer(ByteBuffbuffer)
@@ -498,21 +498,21 @@
void
-FileIOEngine.write(ByteBuffsrcBuffer,
+ByteBufferIOEngine.write(ByteBuffsrcBuffer,
longoffset)
void
+FileIOEngine.write(ByteBuffsrcBuffer,
+ longoffset)
+
+
+void
IOEngine.write(ByteBuffsrcBuffer,
longoffset)
Transfers the data from the given MultiByteBuffer to
IOEngine
-
-void
-ByteBufferIOEngine.write(ByteBuffsrcBuffer,
- longoffset)
-
void
FileMmapEngine.write(ByteBuffsrcBuffer,
@@ -812,15 +812,6 @@
intindex)
-MultiByteBuff
-MultiByteBuff.put(intoffset,
- ByteBuffsrc,
- intsrcOffset,
- intlength)
-Copies from a src MBB to this MBB.
-
-
-
abstract ByteBuff
ByteBuff.put(intoffset,
ByteBuffsrc,
@@ -829,13 +820,22 @@
Copies the contents from the src ByteBuff to this
ByteBuff.
-
+
SingleByteBuff
SingleByteBuff.put(intoffset,
ByteBuffsrc,
intsrcOffset,
intlength)
+
+MultiByteBuff
+MultiByteBuff.put(intoffset,
+ ByteBuffsrc,
+ intsrcOffset,
+ intlength)
+Copies from a src MBB to this MBB.
+
+
static int
ByteBuff.readCompressedInt(ByteBuffbuf)
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0ab8335e/devapidocs/org/apache/hadoop/hbase/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/package-tree.html
b/devapidocs/org/apache/hadoop/hbase/package-tree.html
index 96418c6..36170ca 100644
--- a/devapidocs/org/apache/hadoop/hbase/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/package-tree.html
@@ -445,20 +445,20 @@
java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
title="class or interface in java.io">Serializable)
-org.apache.hadoop.hbase.ClusterMetrics.Option
-org.apache.hadoop.hbase.ProcedureState
org.apache.hadoop.hbase.KeyValue.Type
+org.apache.hadoop.hbase.Size.Unit
org.apache.hadoop.hbase.CellBuilderType
org.apache.hadoop.hbase.HealthChecker.HealthCheckerExitStatus
+org.apache.hadoop.hbase.CompatibilitySingletonFactory.SingletonStorage
org.apache.hadoop.hbase.MetaTableAccessor.QueryType
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
index 6a0cad6..6fcf3a7 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironment.html
@@ -158,7 +158,7 @@
150// only one MasterCoprocessorHost
instance in the master process
151boolean coprocessorsEnabled =
conf.getBoolean(COPROCESSORS_ENABLED_CONF_KEY,
152 DEFAULT_COPROCESSORS_ENABLED);
-153LOG.info("System coprocessor loading
is " + (coprocessorsEnabled ? "enabled" : "disabled"));
+153LOG.trace("System coprocessor loading
is {}", (coprocessorsEnabled ? "enabled" : "disabled"));
154loadSystemCoprocessors(conf,
MASTER_COPROCESSOR_CONF_KEY);
155 }
156
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironmentForCoreCoprocessors.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironmentForCoreCoprocessors.html
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironmentForCoreCoprocessors.html
index 6a0cad6..6fcf3a7 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironmentForCoreCoprocessors.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterEnvironmentForCoreCoprocessors.html
@@ -158,7 +158,7 @@
150// only one MasterCoprocessorHost
instance in the master process
151boolean coprocessorsEnabled =
conf.getBoolean(COPROCESSORS_ENABLED_CONF_KEY,
152 DEFAULT_COPROCESSORS_ENABLED);
-153LOG.info("System coprocessor loading
is " + (coprocessorsEnabled ? "enabled" : "disabled"));
+153LOG.trace("System coprocessor loading
is {}", (coprocessorsEnabled ? "enabled" : "disabled"));
154loadSystemCoprocessors(conf,
MASTER_COPROCESSOR_CONF_KEY);
155 }
156
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html
index 6a0cad6..6fcf3a7 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.MasterObserverOperation.html
@@ -158,7 +158,7 @@
150// only one MasterCoprocessorHost
instance in the master process
151boolean coprocessorsEnabled =
conf.getBoolean(COPROCESSORS_ENABLED_CONF_KEY,
152 DEFAULT_COPROCESSORS_ENABLED);
-153LOG.info("System coprocessor loading
is " + (coprocessorsEnabled ? "enabled" : "disabled"));
+153LOG.trace("System coprocessor loading
is {}", (coprocessorsEnabled ? "enabled" : "disabled"));
154loadSystemCoprocessors(conf,
MASTER_COPROCESSOR_CONF_KEY);
155 }
156
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html
index 6a0cad6..6fcf3a7 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/master/MasterCoprocessorHost.html
@@ -158,7 +158,7 @@
150// only one MasterCoprocessorHost
instance in the master process
151boolean coprocessorsEnabled =
conf.getBoolean(COPROCESSORS_ENABLED_CONF_KEY,
152 DEFAULT_COPROCESSORS_ENABLED);
-153LOG.info("System coprocessor loading
is " + (coprocessorsEnabled ? "enabled" : "disabled"));
+153LOG.trace("System coprocessor loading
is {}", (coprocessorsEnabled ? "enabled" : "disabled"));
154loadSystemCoprocessors(conf,
MASTER_COPROCESSOR_CONF_KEY);
155 }
156
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/6674e3ab/devapidocs/src-html/org/apache/hadoop/hbase/master/ServerManager.html
--
diff --git
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1f2eeb22/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
index 04bf42c..70c0a84 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/master/assignment/MergeTableRegionsProcedure.html
@@ -271,527 +271,533 @@
263return Flow.HAS_MORE_STATE;
264 }
265
-266 @Override
-267 protected void rollbackState(
-268 final MasterProcedureEnv env,
-269 final MergeTableRegionsState state)
throws IOException, InterruptedException {
-270if (isTraceEnabled()) {
-271 LOG.trace(this + " rollback state="
+ state);
-272}
-273
-274try {
-275 switch (state) {
-276 case
MERGE_TABLE_REGIONS_POST_OPERATION:
-277 case
MERGE_TABLE_REGIONS_OPEN_MERGED_REGION:
-278 case
MERGE_TABLE_REGIONS_POST_MERGE_COMMIT_OPERATION:
-279 case
MERGE_TABLE_REGIONS_UPDATE_META:
-280String msg = this + " We are in
the " + state + " state."
-281+ " It is complicated to
rollback the merge operation that region server is working on."
-282+ " Rollback is not supported
and we should let the merge operation to complete";
-283LOG.warn(msg);
-284// PONR
-285throw new
UnsupportedOperationException(this + " unhandled state=" + state);
-286 case
MERGE_TABLE_REGIONS_PRE_MERGE_COMMIT_OPERATION:
-287break;
-288 case
MERGE_TABLE_REGIONS_CREATE_MERGED_REGION:
-289cleanupMergedRegion(env);
-290break;
-291 case
MERGE_TABLE_REGIONS_CLOSE_REGIONS:
-292
rollbackCloseRegionsForMerge(env);
+266 /**
+267 * To rollback {@link
MergeTableRegionsProcedure}, two AssignProcedures are asynchronously
+268 * submitted for each region to be
merged (rollback doesn't wait on the completion of the
+269 * AssignProcedures) . This can be
improved by changing rollback() to support sub-procedures.
+270 * See HBASE-19851 for details.
+271 */
+272 @Override
+273 protected void rollbackState(
+274 final MasterProcedureEnv env,
+275 final MergeTableRegionsState state)
throws IOException, InterruptedException {
+276if (isTraceEnabled()) {
+277 LOG.trace(this + " rollback state="
+ state);
+278}
+279
+280try {
+281 switch (state) {
+282 case
MERGE_TABLE_REGIONS_POST_OPERATION:
+283 case
MERGE_TABLE_REGIONS_OPEN_MERGED_REGION:
+284 case
MERGE_TABLE_REGIONS_POST_MERGE_COMMIT_OPERATION:
+285 case
MERGE_TABLE_REGIONS_UPDATE_META:
+286String msg = this + " We are in
the " + state + " state."
+287+ " It is complicated to
rollback the merge operation that region server is working on."
+288+ " Rollback is not supported
and we should let the merge operation to complete";
+289LOG.warn(msg);
+290// PONR
+291throw new
UnsupportedOperationException(this + " unhandled state=" + state);
+292 case
MERGE_TABLE_REGIONS_PRE_MERGE_COMMIT_OPERATION:
293break;
-294 case
MERGE_TABLE_REGIONS_PRE_MERGE_OPERATION:
-295postRollBackMergeRegions(env);
+294 case
MERGE_TABLE_REGIONS_CREATE_MERGED_REGION:
+295cleanupMergedRegion(env);
296break;
-297 case MERGE_TABLE_REGIONS_PREPARE:
-298break;
-299 default:
-300throw new
UnsupportedOperationException(this + " unhandled state=" + state);
-301 }
-302} catch (Exception e) {
-303 // This will be retried. Unless
there is a bug in the code,
-304 // this should be just a "temporary
error" (e.g. network down)
-305 LOG.warn("Failed rollback attempt
step " + state + " for merging the regions "
-306 +
RegionInfo.getShortNameToLog(regionsToMerge) + " in table " + getTableName(),
e);
-307 throw e;
-308}
-309 }
-310
-311 /*
-312 * Check whether we are in the state
that can be rollback
-313 */
-314 @Override
-315 protected boolean
isRollbackSupported(final MergeTableRegionsState state) {
-316switch (state) {
-317case
MERGE_TABLE_REGIONS_POST_OPERATION:
-318case
MERGE_TABLE_REGIONS_OPEN_MERGED_REGION:
-319case
MERGE_TABLE_REGIONS_POST_MERGE_COMMIT_OPERATION:
-320case
MERGE_TABLE_REGIONS_UPDATE_META:
-321// It is not safe to rollback if
we reach to these states.
-322return false;
-323 default:
-324break;
-325}
-326return true;
-327 }
-328
-329 @Override
-330 protected MergeTableRegionsState
getState(final int stateId) {
-331return
MergeTableRegionsState.forNumber(stateId);
-332 }
-333
-334 @Override
-335 protected
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/client/TestClientPushback.html
--
diff --git
a/testdevapidocs/org/apache/hadoop/hbase/client/TestClientPushback.html
b/testdevapidocs/org/apache/hadoop/hbase/client/TestClientPushback.html
index ed294c6..4ba91bb 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/client/TestClientPushback.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/client/TestClientPushback.html
@@ -109,7 +109,7 @@ var activeTableTab = "activeTableTab";
-public class TestClientPushback
+public class TestClientPushback
extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
Test that we can actually send and use region metrics to
slowdown client writes
@@ -131,26 +131,30 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
Field and Description
+static HBaseClassTestRule
+CLASS_RULE
+
+
private static byte[]
family
-
+
private static long
flushSizeBytes
-
+
private static org.slf4j.Logger
LOG
-
+
private static byte[]
qualifier
-
+
private static
org.apache.hadoop.hbase.TableName
tableName
-
+
private static HBaseTestingUtility
UTIL
@@ -224,13 +228,22 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
Field Detail
+
+
+
+
+
+CLASS_RULE
+public static finalHBaseClassTestRule CLASS_RULE
+
+
LOG
-private static finalorg.slf4j.Logger LOG
+private static finalorg.slf4j.Logger LOG
@@ -239,7 +252,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
UTIL
-private static finalHBaseTestingUtility UTIL
+private static finalHBaseTestingUtility UTIL
@@ -248,7 +261,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
tableName
-private static finalorg.apache.hadoop.hbase.TableName tableName
+private static finalorg.apache.hadoop.hbase.TableName tableName
@@ -257,7 +270,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
family
-private static finalbyte[] family
+private static finalbyte[] family
@@ -266,7 +279,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
qualifier
-private static finalbyte[] qualifier
+private static finalbyte[] qualifier
@@ -275,7 +288,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
flushSizeBytes
-private static finallong flushSizeBytes
+private static finallong flushSizeBytes
See Also:
Constant
Field Values
@@ -296,7 +309,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
TestClientPushback
-publicTestClientPushback()
+publicTestClientPushback()
@@ -313,7 +326,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
setupCluster
-public staticvoidsetupCluster()
+public staticvoidsetupCluster()
throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
title="class or interface in java.lang">Exception
Throws:
@@ -327,7 +340,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
teardownCluster
-public staticvoidteardownCluster()
+public staticvoidteardownCluster()
throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
title="class or interface in java.lang">Exception
Throws:
@@ -341,7 +354,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
testClientTracksServerPushback
-publicvoidtestClientTracksServerPushback()
+publicvoidtestClientTracksServerPushback()
throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true;
title="class or interface in java.lang">Exception
Throws:
@@ -355,7 +368,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
testMutateRowStats
-publicvoidtestMutateRowStats()
+publicvoidtestMutateRowStats()
throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
Throws:
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/cc6597ec/testdevapidocs/org/apache/hadoop/hbase/client/TestClientScanner.MockClientScanner.html
--
diff --git
a/testdevapidocs/org/apache/hadoop/hbase/client/TestClientScanner.MockClientScanner.html
b/testdevapidocs/org/apache/hadoop/hbase/client/TestClientScanner.MockClientScanner.html
index 38aa700..a0fc356 100644
---
a/testdevapidocs/org/apache/hadoop/hbase/client/TestClientScanner.MockClientScanner.html
+++
b/testdevapidocs/org/apache/hadoop/hbase/client/TestClientScanner.MockClientScanner.html
@@ -132,7 +132,7 @@ var
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/aa7ffc92/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupMergeJob.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupMergeJob.html
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupMergeJob.html
index e96923b..67775bf 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupMergeJob.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupMergeJob.html
@@ -46,287 +46,272 @@
038import
org.apache.hadoop.hbase.backup.impl.BackupManifest;
039import
org.apache.hadoop.hbase.backup.impl.BackupSystemTable;
040import
org.apache.hadoop.hbase.backup.util.BackupUtils;
-041import
org.apache.yetus.audience.InterfaceAudience;
-042import org.slf4j.Logger;
-043import org.slf4j.LoggerFactory;
-044import
org.apache.hadoop.hbase.client.Connection;
-045import
org.apache.hadoop.hbase.client.ConnectionFactory;
-046import
org.apache.hadoop.hbase.util.Pair;
-047import org.apache.hadoop.util.Tool;
+041import
org.apache.hadoop.hbase.client.Connection;
+042import
org.apache.hadoop.hbase.client.ConnectionFactory;
+043import
org.apache.hadoop.hbase.util.Pair;
+044import org.apache.hadoop.util.Tool;
+045import
org.apache.yetus.audience.InterfaceAudience;
+046import org.slf4j.Logger;
+047import org.slf4j.LoggerFactory;
048
049/**
050 * MapReduce implementation of {@link
BackupMergeJob}
051 * Must be initialized with configuration
of a backup destination cluster
052 *
053 */
-054
-055@InterfaceAudience.Private
-056public class MapReduceBackupMergeJob
implements BackupMergeJob {
-057 public static final Logger LOG =
LoggerFactory.getLogger(MapReduceBackupMergeJob.class);
-058
-059 protected Tool player;
-060 protected Configuration conf;
-061
-062 public MapReduceBackupMergeJob() {
-063 }
-064
-065 @Override
-066 public Configuration getConf() {
-067return conf;
-068 }
-069
-070 @Override
-071 public void setConf(Configuration conf)
{
-072this.conf = conf;
-073 }
-074
-075 @Override
-076 public void run(String[] backupIds)
throws IOException {
-077String bulkOutputConfKey;
-078
-079// TODO : run player on remote
cluster
-080player = new
MapReduceHFileSplitterJob();
-081bulkOutputConfKey =
MapReduceHFileSplitterJob.BULK_OUTPUT_CONF_KEY;
-082// Player reads all files in
arbitrary directory structure and creates
-083// a Map task for each file
-084String bids =
StringUtils.join(backupIds, ",");
-085
-086if (LOG.isDebugEnabled()) {
-087 LOG.debug("Merge backup images " +
bids);
-088}
-089
-090ListPairTableName,
Path processedTableList = new ArrayListPairTableName,
Path();
-091boolean finishedTables = false;
-092Connection conn =
ConnectionFactory.createConnection(getConf());
-093BackupSystemTable table = new
BackupSystemTable(conn);
-094FileSystem fs =
FileSystem.get(getConf());
-095
-096try {
-097
-098 // Get exclusive lock on backup
system
-099
table.startBackupExclusiveOperation();
-100 // Start merge operation
-101
table.startMergeOperation(backupIds);
-102
-103 // Select most recent backup id
-104 String mergedBackupId =
findMostRecentBackupId(backupIds);
-105
-106 TableName[] tableNames =
getTableNamesInBackupImages(backupIds);
-107 String backupRoot = null;
-108
-109 BackupInfo bInfo =
table.readBackupInfo(backupIds[0]);
-110 backupRoot =
bInfo.getBackupRootDir();
-111
-112 for (int i = 0; i
tableNames.length; i++) {
-113
-114LOG.info("Merge backup images for
" + tableNames[i]);
-115
-116// Find input directories for
table
-117
-118Path[] dirPaths =
findInputDirectories(fs, backupRoot, tableNames[i], backupIds);
-119String dirs =
StringUtils.join(dirPaths, ",");
-120Path bulkOutputPath =
-121
BackupUtils.getBulkOutputDir(BackupUtils.getFileNameCompatibleString(tableNames[i]),
-122 getConf(), false);
-123// Delete content if exists
-124if (fs.exists(bulkOutputPath))
{
-125 if (!fs.delete(bulkOutputPath,
true)) {
-126LOG.warn("Can not delete: " +
bulkOutputPath);
-127 }
-128}
-129Configuration conf = getConf();
-130conf.set(bulkOutputConfKey,
bulkOutputPath.toString());
-131String[] playerArgs = { dirs,
tableNames[i].getNameAsString() };
-132
-133int result = 0;
-134
-135player.setConf(getConf());
-136result =
player.run(playerArgs);
-137if (!succeeded(result)) {
-138 throw new IOException("Can not
merge backup images for " + dirs
-139 + " (check Hadoop/MR and
HBase logs). Player return code =" + result);
-140}
-141// Add to processed table list
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/96e5e102/testdevapidocs/overview-tree.html
--
diff --git a/testdevapidocs/overview-tree.html
b/testdevapidocs/overview-tree.html
index 97691f1..ac787b0 100644
--- a/testdevapidocs/overview-tree.html
+++ b/testdevapidocs/overview-tree.html
@@ -545,24 +545,6 @@
org.apache.hadoop.hbase.io.TestImmutableBytesWritable
org.apache.hadoop.hbase.TestKeyValue
org.apache.hadoop.hbase.regionserver.TestKeyValueScanFixture
-org.apache.hadoop.hbase.rest.model.TestModelBaseT
-
-org.apache.hadoop.hbase.rest.model.TestCellModel
-org.apache.hadoop.hbase.rest.model.TestCellSetModel
-org.apache.hadoop.hbase.rest.model.TestColumnSchemaModel
-org.apache.hadoop.hbase.rest.model.TestNamespacesInstanceModel
-org.apache.hadoop.hbase.rest.model.TestNamespacesModel
-org.apache.hadoop.hbase.rest.model.TestRowModel
-org.apache.hadoop.hbase.rest.model.TestScannerModel
-org.apache.hadoop.hbase.rest.model.TestStorageClusterStatusModel
-org.apache.hadoop.hbase.rest.model.TestStorageClusterVersionModel
-org.apache.hadoop.hbase.rest.model.TestTableInfoModel
-org.apache.hadoop.hbase.rest.model.TestTableListModel
-org.apache.hadoop.hbase.rest.model.TestTableRegionModel
-org.apache.hadoop.hbase.rest.model.TestTableSchemaModel
-org.apache.hadoop.hbase.rest.model.TestVersionModel
-
-
org.apache.hadoop.hbase.util.TestPoolMap.TestPoolType
org.apache.hadoop.hbase.util.TestPoolMap.TestReusablePoolType
@@ -970,7 +952,8 @@
org.apache.hadoop.hbase.filter.TestFilterList.TransformFilter
org.apache.hadoop.hbase.filter.TestInvocationRecordFilter.InvocationRecordFilter
org.apache.hadoop.hbase.coprocessor.TestRegionObserverScannerOpenHook.NoDataFilter
-org.apache.hadoop.hbase.regionserver.TestScannerHeartbeatMessages.SparseFilter
+org.apache.hadoop.hbase.regionserver.TestScannerHeartbeatMessages.SparseCellFilter
+org.apache.hadoop.hbase.regionserver.TestScannerHeartbeatMessages.SparseRowFilter
org.apache.hadoop.hbase.regionserver.querymatcher.TestUserScanQueryMatcher.AlwaysIncludeAndSeekNextRowFilter
org.apache.hadoop.hbase.regionserver.querymatcher.TestUserScanQueryMatcher.AlwaysIncludeFilter
@@ -2915,6 +2898,24 @@
org.apache.hadoop.hbase.mob.TestMobFileName
org.apache.hadoop.hbase.regionserver.TestMobStoreCompaction
org.apache.hadoop.hbase.regionserver.TestMobStoreScanner
+org.apache.hadoop.hbase.rest.model.TestModelBaseT
+
+org.apache.hadoop.hbase.rest.model.TestCellModel
+org.apache.hadoop.hbase.rest.model.TestCellSetModel
+org.apache.hadoop.hbase.rest.model.TestColumnSchemaModel
+org.apache.hadoop.hbase.rest.model.TestNamespacesInstanceModel
+org.apache.hadoop.hbase.rest.model.TestNamespacesModel
+org.apache.hadoop.hbase.rest.model.TestRowModel
+org.apache.hadoop.hbase.rest.model.TestScannerModel
+org.apache.hadoop.hbase.rest.model.TestStorageClusterStatusModel
+org.apache.hadoop.hbase.rest.model.TestStorageClusterVersionModel
+org.apache.hadoop.hbase.rest.model.TestTableInfoModel
+org.apache.hadoop.hbase.rest.model.TestTableListModel
+org.apache.hadoop.hbase.rest.model.TestTableRegionModel
+org.apache.hadoop.hbase.rest.model.TestTableSchemaModel
+org.apache.hadoop.hbase.rest.model.TestVersionModel
+
+
org.apache.hadoop.hbase.master.procedure.TestModifyNamespaceProcedure
org.apache.hadoop.hbase.TestMovedRegionsCleaner
org.apache.hadoop.hbase.nio.TestMultiByteBuff
@@ -3860,7 +3861,6 @@
org.apache.hadoop.hbase.quotas.TestSnapshotQuotaObserverChore.IOThrowingRunnable
org.apache.hadoop.hbase.master.TestSplitLogManager.Expr
org.apache.hadoop.hbase.rest.TestTableScan.ClientSideCellSetModel.Listener
-org.apache.hadoop.hbase.security.token.TestTokenAuthentication.AuthenticationServiceSecurityInfo
org.apache.hadoop.hbase.TimestampTestBase.FlushCache
org.apache.hadoop.hbase.testclassification.VerySlowMapReduceTests
org.apache.hadoop.hbase.testclassification.VerySlowRegionServerTests
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/96e5e102/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestRawAsyncScanCursor.html
--
diff --git
a/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestRawAsyncScanCursor.html
b/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestRawAsyncScanCursor.html
index 5c921e5..6009630 100644
---
a/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestRawAsyncScanCursor.html
+++
b/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestRawAsyncScanCursor.html
@@ -28,150 +28,148 @@
020import static
org.junit.Assert.assertArrayEquals;
021import static
org.junit.Assert.assertEquals;
022import static
org.junit.Assert.assertFalse;
-023
-024import java.io.IOException;
-025import
java.util.concurrent.CompletableFuture;
-026import
java.util.concurrent.ExecutionException;
-027
-028import
org.apache.hadoop.hbase.testclassification.ClientTests;
-029import
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8118541f/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionFileSystem.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionFileSystem.html
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionFileSystem.html
index 6e2a4d4..3931484 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionFileSystem.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionFileSystem.html
@@ -182,7 +182,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
-private static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
+(package private) static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
REGION_TEMP_DIR
Temporary subdirectory of the region directory used for
compaction output.
@@ -708,7 +708,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
REGION_TEMP_DIR
-private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String REGION_TEMP_DIR
+static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String REGION_TEMP_DIR
Temporary subdirectory of the region directory used for
compaction output.
See Also:
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8118541f/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
index cd074b5..e5f8df7 100644
---
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
+++
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.CompactionChecker.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
-private static class HRegionServer.CompactionChecker
+private static class HRegionServer.CompactionChecker
extends ScheduledChore
@@ -233,7 +233,7 @@ extends
instance
-private finalHRegionServer instance
+private finalHRegionServer instance
@@ -242,7 +242,7 @@ extends
majorCompactPriority
-private finalint majorCompactPriority
+private finalint majorCompactPriority
@@ -251,7 +251,7 @@ extends
DEFAULT_PRIORITY
-private static finalint DEFAULT_PRIORITY
+private static finalint DEFAULT_PRIORITY
See Also:
Constant
Field Values
@@ -264,7 +264,7 @@ extends
iteration
-privatelong iteration
+privatelong iteration
@@ -281,7 +281,7 @@ extends
CompactionChecker
-CompactionChecker(HRegionServerh,
+CompactionChecker(HRegionServerh,
intsleepTime,
Stoppablestopper)
@@ -300,7 +300,7 @@ extends
chore
-protectedvoidchore()
+protectedvoidchore()
Description copied from
class:ScheduledChore
The task to execute on each scheduled execution of the
Chore
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8118541f/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
index a387479..983e481 100644
---
a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
+++
b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegionServer.MovedRegionInfo.html
@@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab";
-private static class HRegionServer.MovedRegionInfo
+private static class HRegionServer.MovedRegionInfo
extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
@@ -218,7 +218,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
serverName
-private finalServerName serverName
+private finalServerName serverName
@@ -227,7 +227,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
seqNum
-private finallong seqNum
+private finallong seqNum
@@ -236,7 +236,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
ts
-private finallong ts
+private finallong ts
@@ -253,7 +253,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
MovedRegionInfo
-publicMovedRegionInfo(ServerNameserverName,
+publicMovedRegionInfo(ServerNameserverName,
longcloseSeqNum)
@@ -271,7 +271,7 @@
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/14db89d7/testdevapidocs/src-html/org/apache/hadoop/hbase/master/assignment/TestAssignmentManager.MockRSExecutor.html
--
diff --git
a/testdevapidocs/src-html/org/apache/hadoop/hbase/master/assignment/TestAssignmentManager.MockRSExecutor.html
b/testdevapidocs/src-html/org/apache/hadoop/hbase/master/assignment/TestAssignmentManager.MockRSExecutor.html
index f1db5ca..d8515d7 100644
---
a/testdevapidocs/src-html/org/apache/hadoop/hbase/master/assignment/TestAssignmentManager.MockRSExecutor.html
+++
b/testdevapidocs/src-html/org/apache/hadoop/hbase/master/assignment/TestAssignmentManager.MockRSExecutor.html
@@ -32,813 +32,820 @@
024import static org.junit.Assert.fail;
025
026import java.io.IOException;
-027import java.net.SocketTimeoutException;
-028import java.util.NavigableMap;
-029import java.util.Random;
-030import java.util.Set;
-031import java.util.SortedSet;
-032import
java.util.concurrent.ConcurrentSkipListMap;
-033import
java.util.concurrent.ConcurrentSkipListSet;
-034import
java.util.concurrent.ExecutionException;
-035import java.util.concurrent.Executors;
-036import java.util.concurrent.Future;
-037import
java.util.concurrent.ScheduledExecutorService;
-038import java.util.concurrent.TimeUnit;
-039
-040import
org.apache.hadoop.conf.Configuration;
-041import
org.apache.hadoop.hbase.CategoryBasedTimeout;
-042import
org.apache.hadoop.hbase.DoNotRetryIOException;
-043import
org.apache.hadoop.hbase.HBaseTestingUtility;
-044import
org.apache.hadoop.hbase.NotServingRegionException;
-045import
org.apache.hadoop.hbase.ServerName;
-046import
org.apache.hadoop.hbase.TableName;
-047import
org.apache.hadoop.hbase.client.RegionInfo;
-048import
org.apache.hadoop.hbase.client.RegionInfoBuilder;
-049import
org.apache.hadoop.hbase.client.RetriesExhaustedException;
-050import
org.apache.hadoop.hbase.exceptions.UnexpectedStateException;
-051import
org.apache.hadoop.hbase.ipc.ServerNotRunningYetException;
-052import
org.apache.hadoop.hbase.master.MasterServices;
-053import
org.apache.hadoop.hbase.master.RegionState.State;
-054import
org.apache.hadoop.hbase.master.procedure.MasterProcedureConstants;
-055import
org.apache.hadoop.hbase.master.procedure.ProcedureSyncWait;
-056import
org.apache.hadoop.hbase.master.procedure.RSProcedureDispatcher;
-057import
org.apache.hadoop.hbase.procedure2.Procedure;
-058import
org.apache.hadoop.hbase.procedure2.ProcedureMetrics;
-059import
org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility;
-060import
org.apache.hadoop.hbase.procedure2.store.wal.WALProcedureStore;
-061import
org.apache.hadoop.hbase.procedure2.util.StringUtils;
-062import
org.apache.hadoop.hbase.regionserver.RegionServerAbortedException;
-063import
org.apache.hadoop.hbase.regionserver.RegionServerStoppedException;
-064import
org.apache.hadoop.hbase.testclassification.MasterTests;
-065import
org.apache.hadoop.hbase.testclassification.MediumTests;
-066import
org.apache.hadoop.hbase.util.Bytes;
-067import
org.apache.hadoop.hbase.util.FSUtils;
-068import
org.apache.hadoop.ipc.RemoteException;
-069import org.junit.After;
-070import org.junit.Before;
-071import org.junit.Ignore;
-072import org.junit.Rule;
-073import org.junit.Test;
-074import
org.junit.experimental.categories.Category;
-075import
org.junit.rules.ExpectedException;
-076import org.junit.rules.TestName;
-077import org.junit.rules.TestRule;
-078import org.slf4j.Logger;
-079import org.slf4j.LoggerFactory;
-080import
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-081import
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionRequest;
-082import
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.CloseRegionResponse;
-083import
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ExecuteProceduresRequest;
-084import
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.ExecuteProceduresResponse;
-085import
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionRequest;
-086import
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionRequest.RegionOpenInfo;
-087import
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionResponse;
-088import
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionResponse.RegionOpeningState;
-089import
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition;
-090import
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.RegionStateTransition.TransitionCode;
-091import
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.ReportRegionStateTransitionRequest;
-092
-093@Category({MasterTests.class,
MediumTests.class})
-094public class TestAssignmentManager {
-095 private static final Logger LOG =
LoggerFactory.getLogger(TestAssignmentManager.class);
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/0b638133/testdevapidocs/src-html/org/apache/hadoop/hbase/TestCellUtil.TestCellScanner.html
--
diff --git
a/testdevapidocs/src-html/org/apache/hadoop/hbase/TestCellUtil.TestCellScanner.html
b/testdevapidocs/src-html/org/apache/hadoop/hbase/TestCellUtil.TestCellScanner.html
index 232ef56..bc3a6d0 100644
---
a/testdevapidocs/src-html/org/apache/hadoop/hbase/TestCellUtil.TestCellScanner.html
+++
b/testdevapidocs/src-html/org/apache/hadoop/hbase/TestCellUtil.TestCellScanner.html
@@ -29,610 +29,626 @@
021import static
org.junit.Assert.assertEquals;
022import static
org.junit.Assert.assertFalse;
023import static
org.junit.Assert.assertTrue;
-024
-025import java.io.ByteArrayOutputStream;
-026import java.io.IOException;
-027import java.math.BigDecimal;
-028import java.nio.ByteBuffer;
-029import java.util.ArrayList;
-030import java.util.List;
-031import java.util.NavigableMap;
-032import java.util.TreeMap;
-033import
org.apache.hadoop.hbase.testclassification.MiscTests;
-034import
org.apache.hadoop.hbase.testclassification.SmallTests;
-035import
org.apache.hadoop.hbase.util.Bytes;
-036import org.junit.Assert;
-037import org.junit.Test;
-038import
org.junit.experimental.categories.Category;
-039
-040@Category({MiscTests.class,
SmallTests.class})
-041public class TestCellUtil {
-042 /**
-043 * CellScannable used in test. Returns
a {@link TestCellScanner}
-044 */
-045 private static class TestCellScannable
implements CellScannable {
-046private final int cellsCount;
-047TestCellScannable(final int
cellsCount) {
-048 this.cellsCount = cellsCount;
-049}
-050@Override
-051public CellScanner cellScanner() {
-052 return new
TestCellScanner(this.cellsCount);
-053}
-054 }
-055
-056 /**
-057 * CellScanner used in test.
-058 */
-059 private static class TestCellScanner
implements CellScanner {
-060private int count = 0;
-061private Cell current = null;
-062private final int cellsCount;
-063
-064TestCellScanner(final int cellsCount)
{
-065 this.cellsCount = cellsCount;
-066}
-067
-068@Override
-069public Cell current() {
-070 return this.current;
-071}
-072
-073@Override
-074public boolean advance() throws
IOException {
-075 if (this.count cellsCount) {
-076this.current = new
TestCell(this.count);
-077this.count++;
-078return true;
-079 }
-080 return false;
-081}
-082 }
-083
-084 /**
-085 * Cell used in test. Has row only.
-086 */
-087 private static class TestCell
implements Cell {
-088private final byte [] row;
-089
-090TestCell(final int i) {
-091 this.row = Bytes.toBytes(i);
-092}
-093
-094@Override
-095public byte[] getRowArray() {
-096 return this.row;
-097}
-098
-099@Override
-100public int getRowOffset() {
-101 return 0;
-102}
-103
-104@Override
-105public short getRowLength() {
-106 return (short)this.row.length;
-107}
-108
-109@Override
-110public byte[] getFamilyArray() {
-111 // TODO Auto-generated method
stub
-112 return null;
-113}
-114
-115@Override
-116public int getFamilyOffset() {
-117 // TODO Auto-generated method
stub
-118 return 0;
-119}
-120
-121@Override
-122public byte getFamilyLength() {
-123 // TODO Auto-generated method
stub
-124 return 0;
-125}
-126
-127@Override
-128public byte[] getQualifierArray() {
-129 // TODO Auto-generated method
stub
-130 return null;
-131}
-132
-133@Override
-134public int getQualifierOffset() {
-135 // TODO Auto-generated method
stub
-136 return 0;
-137}
-138
-139@Override
-140public int getQualifierLength() {
-141 // TODO Auto-generated method
stub
-142 return 0;
-143}
-144
-145@Override
-146public long getTimestamp() {
-147 // TODO Auto-generated method
stub
-148 return 0;
-149}
-150
-151@Override
-152public byte getTypeByte() {
-153 // TODO Auto-generated method
stub
-154 return 0;
-155}
-156
-157@Override
-158public byte[] getValueArray() {
-159 // TODO Auto-generated method
stub
-160 return null;
-161}
-162
-163@Override
-164public int getValueOffset() {
-165 // TODO Auto-generated method
stub
-166 return 0;
-167}
-168
-169@Override
-170public int getValueLength() {
-171 // TODO Auto-generated method
stub
-172 return 0;
-173}
-174
-175@Override
-176public byte[] getTagsArray() {
-177 // TODO Auto-generated method
stub
-178 return null;
-179}
-180
-181@Override
-182public int getTagsOffset() {
-183 // TODO Auto-generated method
stub
-184 return 0;
-185}
-186
-187@Override
-188public long getSequenceId() {
-189
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f183e80f/testdevapidocs/org/apache/hadoop/hbase/regionserver/wal/TestDurability.html
--
diff --git
a/testdevapidocs/org/apache/hadoop/hbase/regionserver/wal/TestDurability.html
b/testdevapidocs/org/apache/hadoop/hbase/regionserver/wal/TestDurability.html
index 5f0e47f..8a340d3 100644
---
a/testdevapidocs/org/apache/hadoop/hbase/regionserver/wal/TestDurability.html
+++
b/testdevapidocs/org/apache/hadoop/hbase/regionserver/wal/TestDurability.html
@@ -18,7 +18,7 @@
catch(err) {
}
//-->
-var methods =
{"i0":10,"i1":9,"i2":10,"i3":10,"i4":9,"i5":10,"i6":9,"i7":10,"i8":10,"i9":10,"i10":10};
+var methods =
{"i0":10,"i1":10,"i2":9,"i3":10,"i4":10,"i5":9,"i6":10,"i7":9,"i8":10,"i9":10,"i10":10,"i11":10};
var tabs = {65535:["t0","All Methods"],1:["t1","Static
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
var altColor = "altColor";
var rowColor = "rowColor";
@@ -109,7 +109,7 @@ var activeTableTab = "activeTableTab";
-public class TestDurability
+public class TestDurability
extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
Tests for WAL write durability
@@ -155,14 +155,18 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
FS
+org.junit.rules.TestName
+name
+
+
private static byte[]
ROW
-
+
private static HBaseTestingUtility
TEST_UTIL
-
+
http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
walProvider
@@ -200,48 +204,57 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
private
org.apache.hadoop.hbase.regionserver.HRegion
-createHRegion(byte[]tableName,
- http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">StringcallingMethod,
- org.apache.hadoop.hbase.wal.WALlog,
+createHRegion(org.apache.hadoop.hbase.client.TableDescriptortd,
+ org.apache.hadoop.hbase.client.RegionInfoinfo,
+ http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">Stringdir,
+ org.apache.hadoop.hbase.wal.WALwal,
org.apache.hadoop.hbase.client.Durabilitydurability)
+private
org.apache.hadoop.hbase.regionserver.HRegion
+createHRegion(org.apache.hadoop.hbase.wal.WALFactorywals,
+
org.apache.hadoop.hbase.client.Durabilitydurability)
+
+
static http://docs.oracle.com/javase/8/docs/api/java/lang/Iterable.html?is-external=true;
title="class or interface in java.lang">Iterablehttp://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object[]
data()
-
+
private
org.apache.hadoop.hbase.client.Put
newPut(org.apache.hadoop.hbase.client.Durabilitydurability)
-
+
void
setUp()
-
+
static void
setUpBeforeClass()
-
+
void
tearDown()
-
+
static void
tearDownAfterClass()
-
+
void
testDurability()
-
+
void
testIncrement()
-
+
void
-testIncrementWithReturnResultsSetToFalse()
+testIncrementWithReturnResultsSetToFalse()
+Test when returnResults set to false in increment it should
not return the result instead it
+ resturn null.
+
-
+
private void
verifyWALCount(org.apache.hadoop.hbase.wal.WALFactorywals,
org.apache.hadoop.hbase.wal.WALlog,
@@ -275,7 +288,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
TEST_UTIL
-private static finalHBaseTestingUtility TEST_UTIL
+private static finalHBaseTestingUtility TEST_UTIL
@@ -284,7 +297,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
FS
-private staticorg.apache.hadoop.fs.FileSystem FS
+private staticorg.apache.hadoop.fs.FileSystem FS
@@ -293,7 +306,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
CLUSTER
-private staticorg.apache.hadoop.hdfs.MiniDFSCluster CLUSTER
+private staticorg.apache.hadoop.hdfs.MiniDFSCluster CLUSTER
@@ -302,7 +315,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
CONF
-private staticorg.apache.hadoop.conf.Configuration CONF
+private staticorg.apache.hadoop.conf.Configuration CONF
@@ -311,7 +324,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
DIR
-private staticorg.apache.hadoop.fs.Path DIR
+private staticorg.apache.hadoop.fs.Path DIR
@@ -320,7 +333,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
FAMILY
-private staticbyte[] FAMILY
+private staticbyte[] FAMILY
@@ -329,7 +342,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
ROW
-private staticbyte[] ROW
+private staticbyte[] ROW
@@ -338,16
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureScheduler.FairQueue.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureScheduler.FairQueue.html
b/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureScheduler.FairQueue.html
index 3ab8444..28dc97a 100644
---
a/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureScheduler.FairQueue.html
+++
b/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureScheduler.FairQueue.html
@@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab";
PrevClass
-NextClass
+NextClass
Frames
@@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab";
-private static class MasterProcedureScheduler.FairQueueT
extends http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableT
+private static class MasterProcedureScheduler.FairQueueT
extends http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableT
extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
@@ -244,7 +244,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
quantum
-private finalint quantum
+private finalint quantum
@@ -253,7 +253,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
currentQueue
-privateMasterProcedureScheduler.QueueT extends http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableT currentQueue
+privateMasterProcedureScheduler.QueueT extends http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableT currentQueue
@@ -262,7 +262,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
queueHead
-privateMasterProcedureScheduler.QueueT extends http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableT queueHead
+privateMasterProcedureScheduler.QueueT extends http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true;
title="class or interface in java.lang">ComparableT queueHead
@@ -271,7 +271,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
currentQuantum
-privateint currentQuantum
+privateint currentQuantum
@@ -280,7 +280,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
size
-privateint size
+privateint size
@@ -297,7 +297,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
FairQueue
-publicFairQueue()
+publicFairQueue()
@@ -306,7 +306,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
FairQueue
-publicFairQueue(intquantum)
+publicFairQueue(intquantum)
@@ -323,7 +323,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
hasRunnables
-publicbooleanhasRunnables()
+publicbooleanhasRunnables()
@@ -332,7 +332,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
add
-publicvoidadd(MasterProcedureScheduler.QueueTqueue)
+publicvoidadd(MasterProcedureScheduler.QueueTqueue)
@@ -341,7 +341,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
remove
-publicvoidremove(MasterProcedureScheduler.QueueTqueue)
+publicvoidremove(MasterProcedureScheduler.QueueTqueue)
@@ -350,7 +350,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
poll
-publicMasterProcedureScheduler.QueueTpoll()
+publicMasterProcedureScheduler.QueueTpoll()
@@ -359,7 +359,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
nextQueue
-privatebooleannextQueue()
+privatebooleannextQueue()
@@ -368,7 +368,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
setNextQueue
-privatevoidsetNextQueue(MasterProcedureScheduler.QueueTqueue)
+privatevoidsetNextQueue(MasterProcedureScheduler.QueueTqueue)
@@ -377,7 +377,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
calculateQuantum
-privateintcalculateQuantum(MasterProcedureScheduler.Queuequeue)
+privateintcalculateQuantum(MasterProcedureScheduler.Queuequeue)
@@ -409,7 +409,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
PrevClass
-NextClass
+NextClass
Frames
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/49431b18/devapidocs/org/apache/hadoop/hbase/master/procedure/MasterProcedureScheduler.PeerQueue.html
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c7c40c62/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.html
b/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.html
index 9e7d3ae..c7b483d 100644
--- a/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.html
+++ b/devapidocs/org/apache/hadoop/hbase/security/access/AccessController.html
@@ -18,7 +18,7 @@
catch(err) {
}
//-->
-var methods =
{"i0":9,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":9,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":9,"i25":9,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109"
:10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":10,"i145":10,"i146":10,"i147":10,"i148":10};
+var methods =
{"i0":9,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":9,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":9,"i25":9,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i109"
:10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":10,"i142":10,"i143":10,"i144":10,"i145":10,"i146":10,"i147":10};
var tabs = {65535:["t0","All Methods"],1:["t1","Static
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
var altColor = "altColor";
var rowColor = "rowColor";
@@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab";
@InterfaceAudience.LimitedPrivate(value="Configuration")
-public class AccessController
+public class AccessController
extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
implements MasterCoprocessor, RegionCoprocessor, RegionServerCoprocessor,
org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.Interface,
MasterObserver,
RegionObserver,
RegionServerObserver, EndpointObserver,
BulkLoadObserver
Provides basic authorization checks for data access and
administrative
@@ -959,24 +959,18 @@ implements
void
-preGetClusterStatus(ObserverContextMasterCoprocessorEnvironmentctx)
-Called before get cluster status.
-
-
-
-void
preGetLocks(ObserverContextMasterCoprocessorEnvironmentctx)
Called before a getLocks request has been processed.
-
+
void
preGetNamespaceDescriptor(ObserverContextMasterCoprocessorEnvironmentctx,
http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">Stringnamespace)
Called before a getNamespaceDescriptor request has been
processed.
-
+
void
preGetOp(ObserverContextRegionCoprocessorEnvironmentc,
Getget,
@@ -984,20 +978,20 @@ implements Called before the client performs a Get
-
+
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/bb398572/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/MemcachedBlockCache.HFileBlockTranscoder.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/MemcachedBlockCache.HFileBlockTranscoder.html
b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/MemcachedBlockCache.HFileBlockTranscoder.html
index 8373f27..80df615 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/MemcachedBlockCache.HFileBlockTranscoder.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/io/hfile/MemcachedBlockCache.HFileBlockTranscoder.html
@@ -6,294 +6,293 @@
-001
-002/**
-003 * Copyright The Apache Software
Foundation
-004 *
-005 * Licensed to the Apache Software
Foundation (ASF) under one or more
-006 * contributor license agreements. See
the NOTICE file distributed with this
-007 * work for additional information
regarding copyright ownership. The ASF
-008 * licenses this file to you under the
Apache License, Version 2.0 (the
-009 * "License"); you may not use this file
except in compliance with the License.
-010 * You may obtain a copy of the License
at
-011 *
-012 *
http://www.apache.org/licenses/LICENSE-2.0
-013 *
-014 * Unless required by applicable law or
agreed to in writing, software
-015 * distributed under the License is
distributed on an "AS IS" BASIS, WITHOUT
-016 * WARRANTIES OR CONDITIONS OF ANY KIND,
either express or implied. See the
-017 * License for the specific language
governing permissions and limitations
-018 * under the License.
-019 */
-020
-021package
org.apache.hadoop.hbase.io.hfile;
-022
-023import java.io.IOException;
-024import java.net.InetSocketAddress;
-025import java.nio.ByteBuffer;
-026import java.util.ArrayList;
-027import java.util.Iterator;
-028import java.util.List;
-029import
java.util.NoSuchElementException;
-030import
java.util.concurrent.ExecutionException;
-031
-032import
org.apache.hadoop.conf.Configuration;
-033import
org.apache.hadoop.hbase.HConstants;
-034import
org.apache.hadoop.hbase.io.hfile.Cacheable.MemoryType;
-035import
org.apache.hadoop.hbase.nio.ByteBuff;
-036import
org.apache.hadoop.hbase.nio.SingleByteBuff;
-037import
org.apache.hadoop.hbase.trace.TraceUtil;
-038import
org.apache.hadoop.hbase.util.Addressing;
-039import
org.apache.htrace.core.TraceScope;
-040import
org.apache.yetus.audience.InterfaceAudience;
-041import org.slf4j.Logger;
-042import org.slf4j.LoggerFactory;
-043
-044import net.spy.memcached.CachedData;
-045import
net.spy.memcached.ConnectionFactoryBuilder;
-046import net.spy.memcached.FailureMode;
-047import
net.spy.memcached.MemcachedClient;
-048import
net.spy.memcached.transcoders.Transcoder;
-049
-050/**
-051 * Class to store blocks into
memcached.
-052 * This should only be used on a cluster
of Memcached daemons that are tuned well and have a
-053 * good network connection to the HBase
regionservers. Any other use will likely slow down HBase
-054 * greatly.
-055 */
-056@InterfaceAudience.Private
-057public class MemcachedBlockCache
implements BlockCache {
-058 private static final Logger LOG =
LoggerFactory.getLogger(MemcachedBlockCache.class.getName());
-059
-060 // Some memcache versions won't take
more than 1024 * 1024. So set the limit below
-061 // that just in case this client is
used with those versions.
-062 public static final int MAX_SIZE = 1020
* 1024;
-063
-064 // Config key for what memcached
servers to use.
-065 // They should be specified in a comma
sperated list with ports.
-066 // like:
-067 //
-068 // host1:11211,host3:8080,host4:11211
-069 public static final String
MEMCACHED_CONFIG_KEY = "hbase.cache.memcached.servers";
-070 public static final String
MEMCACHED_TIMEOUT_KEY = "hbase.cache.memcached.timeout";
-071 public static final String
MEMCACHED_OPTIMEOUT_KEY = "hbase.cache.memcached.optimeout";
-072 public static final String
MEMCACHED_OPTIMIZE_KEY = "hbase.cache.memcached.spy.optimze";
-073 public static final long
MEMCACHED_DEFAULT_TIMEOUT = 500;
-074 public static final boolean
MEMCACHED_OPTIMIZE_DEFAULT = false;
-075
-076 private final MemcachedClient client;
-077 private final HFileBlockTranscoder tc =
new HFileBlockTranscoder();
-078 private final CacheStats cacheStats =
new CacheStats("MemcachedBlockCache");
-079
-080 public
MemcachedBlockCache(Configuration c) throws IOException {
-081LOG.info("Creating
MemcachedBlockCache");
-082
-083long opTimeout =
c.getLong(MEMCACHED_OPTIMEOUT_KEY, MEMCACHED_DEFAULT_TIMEOUT);
-084long queueTimeout =
c.getLong(MEMCACHED_TIMEOUT_KEY, opTimeout + MEMCACHED_DEFAULT_TIMEOUT);
-085boolean optimize =
c.getBoolean(MEMCACHED_OPTIMIZE_KEY, MEMCACHED_OPTIMIZE_DEFAULT);
-086
-087ConnectionFactoryBuilder builder =
new ConnectionFactoryBuilder()
-088.setOpTimeout(opTimeout)
-089
.setOpQueueMaxBlockTime(queueTimeout)
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/mapred/class-use/GroupingTableMap.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/GroupingTableMap.html
b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/GroupingTableMap.html
index a591b29..030d8dd 100644
--- a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/GroupingTableMap.html
+++ b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/GroupingTableMap.html
@@ -120,6 +120,6 @@
-Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
+Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/mapred/class-use/HRegionPartitioner.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/HRegionPartitioner.html
b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/HRegionPartitioner.html
index 588d268..a9e5ff5 100644
---
a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/HRegionPartitioner.html
+++
b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/HRegionPartitioner.html
@@ -120,6 +120,6 @@
-Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
+Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableMap.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableMap.html
b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableMap.html
index 5c452b2..cf51754 100644
--- a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableMap.html
+++ b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableMap.html
@@ -120,6 +120,6 @@
-Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
+Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableReduce.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableReduce.html
b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableReduce.html
index 5738603..a5e9992 100644
---
a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableReduce.html
+++
b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/IdentityTableReduce.html
@@ -120,6 +120,6 @@
-Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
+Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/mapred/class-use/MultiTableSnapshotInputFormat.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/MultiTableSnapshotInputFormat.html
b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/MultiTableSnapshotInputFormat.html
index 99e8e24..f29b6e6 100644
---
a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/MultiTableSnapshotInputFormat.html
+++
b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/MultiTableSnapshotInputFormat.html
@@ -120,6 +120,6 @@
-Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
+Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/69506d41/devapidocs/org/apache/hadoop/hbase/mapred/class-use/RowCounter.RowCounterMapper.Counters.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/RowCounter.RowCounterMapper.Counters.html
b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/RowCounter.RowCounterMapper.Counters.html
index 7971d8d..91a5c17 100644
---
a/devapidocs/org/apache/hadoop/hbase/mapred/class-use/RowCounter.RowCounterMapper.Counters.html
+++
b/devapidocs/org/apache/hadoop/hbase/mapred/class-use/RowCounter.RowCounterMapper.Counters.html
@@ -172,6 +172,6 @@ the order they are declared.
-Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
+Copyright 20072018 https://www.apache.org/;>The Apache Software Foundation. All rights
reserved.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/83bf6175/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestAsyncProcess.MyAsyncRequestFutureImpl.html
--
diff --git
a/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestAsyncProcess.MyAsyncRequestFutureImpl.html
b/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestAsyncProcess.MyAsyncRequestFutureImpl.html
index bbd91b8..4f76302 100644
---
a/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestAsyncProcess.MyAsyncRequestFutureImpl.html
+++
b/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestAsyncProcess.MyAsyncRequestFutureImpl.html
@@ -56,1641 +56,1753 @@
048import
java.util.concurrent.atomic.AtomicBoolean;
049import
java.util.concurrent.atomic.AtomicInteger;
050import
java.util.concurrent.atomic.AtomicLong;
-051
-052import
org.apache.hadoop.conf.Configuration;
-053import
org.apache.hadoop.hbase.CallQueueTooBigException;
-054import
org.apache.hadoop.hbase.CategoryBasedTimeout;
-055import org.apache.hadoop.hbase.Cell;
-056import
org.apache.hadoop.hbase.HConstants;
-057import
org.apache.hadoop.hbase.HRegionInfo;
-058import
org.apache.hadoop.hbase.HRegionLocation;
-059import
org.apache.hadoop.hbase.RegionLocations;
-060import
org.apache.hadoop.hbase.ServerName;
-061import
org.apache.hadoop.hbase.TableName;
-062import
org.apache.hadoop.hbase.client.AsyncProcessTask.ListRowAccess;
-063import
org.apache.hadoop.hbase.client.AsyncProcessTask.SubmittedRows;
-064import
org.apache.hadoop.hbase.client.backoff.ClientBackoffPolicy;
-065import
org.apache.hadoop.hbase.client.backoff.ServerStatistics;
-066import
org.apache.hadoop.hbase.client.coprocessor.Batch;
-067import
org.apache.hadoop.hbase.ipc.RpcControllerFactory;
-068import
org.apache.hadoop.hbase.testclassification.ClientTests;
-069import
org.apache.hadoop.hbase.testclassification.MediumTests;
-070import
org.apache.hadoop.hbase.util.Bytes;
-071import
org.apache.hadoop.hbase.util.Threads;
-072import org.junit.Assert;
-073import org.junit.BeforeClass;
-074import org.junit.Ignore;
-075import org.junit.Rule;
-076import org.junit.Test;
-077import
org.junit.experimental.categories.Category;
-078import org.junit.rules.TestRule;
-079import org.mockito.Mockito;
-080import org.slf4j.Logger;
-081import org.slf4j.LoggerFactory;
-082
-083@Category({ClientTests.class,
MediumTests.class})
-084public class TestAsyncProcess {
-085 @Rule public final TestRule timeout =
CategoryBasedTimeout.builder().withTimeout(this.getClass()).
-086
withLookingForStuckThread(true).build();
-087 private static final Logger LOG =
LoggerFactory.getLogger(TestAsyncProcess.class);
-088 private static final TableName
DUMMY_TABLE =
-089 TableName.valueOf("DUMMY_TABLE");
-090 private static final byte[]
DUMMY_BYTES_1 = Bytes.toBytes("DUMMY_BYTES_1");
-091 private static final byte[]
DUMMY_BYTES_2 = Bytes.toBytes("DUMMY_BYTES_2");
-092 private static final byte[]
DUMMY_BYTES_3 = Bytes.toBytes("DUMMY_BYTES_3");
-093 private static final byte[] FAILS =
Bytes.toBytes("FAILS");
-094 private static final Configuration CONF
= new Configuration();
-095 private static final
ConnectionConfiguration CONNECTION_CONFIG =
-096 new
ConnectionConfiguration(CONF);
-097 private static final ServerName sn =
ServerName.valueOf("s1,1,1");
-098 private static final ServerName sn2 =
ServerName.valueOf("s2,2,2");
-099 private static final ServerName sn3 =
ServerName.valueOf("s3,3,3");
-100 private static final HRegionInfo hri1
=
-101 new HRegionInfo(DUMMY_TABLE,
DUMMY_BYTES_1, DUMMY_BYTES_2, false, 1);
-102 private static final HRegionInfo hri2
=
-103 new HRegionInfo(DUMMY_TABLE,
DUMMY_BYTES_2, HConstants.EMPTY_END_ROW, false, 2);
-104 private static final HRegionInfo hri3
=
-105 new HRegionInfo(DUMMY_TABLE,
DUMMY_BYTES_3, HConstants.EMPTY_END_ROW, false, 3);
-106 private static final HRegionLocation
loc1 = new HRegionLocation(hri1, sn);
-107 private static final HRegionLocation
loc2 = new HRegionLocation(hri2, sn);
-108 private static final HRegionLocation
loc3 = new HRegionLocation(hri3, sn2);
-109
-110 // Replica stuff
-111 private static final RegionInfo hri1r1
= RegionReplicaUtil.getRegionInfoForReplica(hri1, 1);
-112 private static final RegionInfo hri1r2
= RegionReplicaUtil.getRegionInfoForReplica(hri1, 2);
-113 private static final RegionInfo hri2r1
= RegionReplicaUtil.getRegionInfoForReplica(hri2, 1);
-114 private static final RegionLocations
hrls1 = new RegionLocations(new HRegionLocation(hri1, sn),
-115 new HRegionLocation(hri1r1, sn2),
new HRegionLocation(hri1r2, sn3));
-116 private static final RegionLocations
hrls2 = new RegionLocations(new HRegionLocation(hri2, sn2),
-117 new HRegionLocation(hri2r1,
sn3));
-118 private static final RegionLocations
hrls3 =
-119 new RegionLocations(new
HRegionLocation(hri3, sn3), null);
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/devapidocs/org/apache/hadoop/hbase/regionserver/CSLMImmutableSegment.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/regionserver/CSLMImmutableSegment.html
b/devapidocs/org/apache/hadoop/hbase/regionserver/CSLMImmutableSegment.html
index b809814..dc68db2 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/CSLMImmutableSegment.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/CSLMImmutableSegment.html
@@ -219,7 +219,7 @@ extends Segment
-close,
compare,
compareRows,
decScannerCount,
dump,
getCellLength,
getCellsCount,
getCellSet,
getComparator,
getFirstAfter,
getMemStoreLAB,
getMinSequenceId,
getScanner,
getScanner,
getScanners,
getTimeRangeTracker,
headSet, heapSize,
heapSizeChange,
incScannerCount,
incSize,
internalAdd,
isEmpty,
isTagsPresent,
iterator, keySize,
last,
maybeCloneWithAllocator,
setCellSet,
shouldSeek,
tailSet,
updateMetaInfo, updateMetaInfo
+close,
compare,
compareRows,
decScannerCount,
dump,
getCellLength,
getCellsCount,
getCellSet,
getComparator,
getFirstAfter,
getMemStoreLAB,
getMinSequenceId,
getScanner,
getScanner,
getScanners,
getTimeRangeTracker,
headSet, heapSize,
heapSizeChange,
incScannerCount,
incSize,
internalAdd,
isEmpty,
isTagsPresent,
iterator, keySize,
last,
maybeCloneWithAllocator,
setCellSet,
shouldSeek,
tailSet,
updateMetaInfo, updateMetaInfo
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/devapidocs/org/apache/hadoop/hbase/regionserver/CellArrayImmutableSegment.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/regionserver/CellArrayImmutableSegment.html
b/devapidocs/org/apache/hadoop/hbase/regionserver/CellArrayImmutableSegment.html
index ad6efee..bbe4de7 100644
---
a/devapidocs/org/apache/hadoop/hbase/regionserver/CellArrayImmutableSegment.html
+++
b/devapidocs/org/apache/hadoop/hbase/regionserver/CellArrayImmutableSegment.html
@@ -248,7 +248,7 @@ extends Segment
-close,
compare,
compareRows,
decScannerCount,
dump,
getCellLength,
getCellsCount,
getCellSet,
getComparator,
getFirstAfter,
getMemStoreLAB,
getMinSequenceId,
getScanner,
getScanner,
getScanners,
getTimeRangeTracker,
headSet, heapSize,
heapSizeChange,
incScannerCount,
incSize,
internalAdd,
isEmpty,
isTagsPresent,
iterator, keySize,
last,
maybeCloneWithAllocator,
setCellSet,
shouldSeek,
tailSet,
updateMetaInfo, updateMetaInfo
+close,
compare,
compareRows,
decScannerCount,
dump,
getCellLength,
getCellsCount,
getCellSet,
getComparator,
getFirstAfter,
getMemStoreLAB,
getMinSequenceId,
getScanner,
getScanner,
getScanners,
getTimeRangeTracker,
headSet, heapSize,
heapSizeChange,
incScannerCount,
incSize,
internalAdd,
isEmpty,
isTagsPresent,
iterator, keySize,
last,
maybeCloneWithAllocator,
setCellSet,
shouldSeek,
tailSet,
updateMetaInfo, updateMetaInfo
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/63d6f712/devapidocs/org/apache/hadoop/hbase/regionserver/CellChunkImmutableSegment.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/regionserver/CellChunkImmutableSegment.html
b/devapidocs/org/apache/hadoop/hbase/regionserver/CellChunkImmutableSegment.html
index 0b339b8..a79894f 100644
---
a/devapidocs/org/apache/hadoop/hbase/regionserver/CellChunkImmutableSegment.html
+++
b/devapidocs/org/apache/hadoop/hbase/regionserver/CellChunkImmutableSegment.html
@@ -18,7 +18,7 @@
catch(err) {
}
//-->
-var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10};
+var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10};
var tabs = {65535:["t0","All Methods"],2:["t2","Instance
Methods"],8:["t4","Concrete Methods"]};
var altColor = "altColor";
var rowColor = "rowColor";
@@ -120,7 +120,7 @@ var activeTableTab = "activeTableTab";
@InterfaceAudience.Private
-public class CellChunkImmutableSegment
+public class CellChunkImmutableSegment
extends ImmutableSegment
CellChunkImmutableSegment extends the API supported by a Segment,
and ImmutableSegment. This
immutable segment is working with CellSet with
@@ -224,22 +224,26 @@ extends canBeFlattened()
+private Cell
+copyCellIntoMSLAB(Cellcell)
+
+
private int
createCellReference(ByteBufferKeyValuecell,
http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true;
title="class or interface in java.nio">ByteBufferidxBuffer,
intidxOffset)
-
+
protected long
indexEntrySize()
-
+
private void
initializeCellSet(intnumOfCells,
MemStoreSegmentsIteratoriterator,
MemStoreCompactionStrategy.Actionaction)
-
+
private void
reinitializeCellSet(intnumOfCells,
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d449e87f/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntry.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntry.html
b/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntry.html
index 19ab2eb..468f000 100644
---
a/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntry.html
+++
b/devapidocs/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.BucketEntry.html
@@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
-static class BucketCache.BucketEntry
+static class BucketCache.BucketEntry
extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true;
title="class or interface in java.io">Serializable
Item in cache. We expect this to be where most memory goes.
Java uses 8
@@ -288,7 +288,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
serialVersionUID
-private static finallong serialVersionUID
+private static finallong serialVersionUID
See Also:
Constant
Field Values
@@ -301,7 +301,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
COMPARATOR
-static finalhttp://docs.oracle.com/javase/8/docs/api/java/util/Comparator.html?is-external=true;
title="class or interface in java.util">ComparatorBucketCache.BucketEntry COMPARATOR
+static finalhttp://docs.oracle.com/javase/8/docs/api/java/util/Comparator.html?is-external=true;
title="class or interface in java.util">ComparatorBucketCache.BucketEntry COMPARATOR
@@ -310,7 +310,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
offsetBase
-privateint offsetBase
+privateint offsetBase
@@ -319,7 +319,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
length
-privateint length
+privateint length
@@ -328,7 +328,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
offset1
-privatebyte offset1
+privatebyte offset1
@@ -337,7 +337,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
deserialiserIndex
-byte deserialiserIndex
+byte deserialiserIndex
@@ -346,7 +346,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
accessCounter
-private volatilelong accessCounter
+private volatilelong accessCounter
@@ -355,7 +355,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
priority
-privateBlockPriority priority
+privateBlockPriority priority
@@ -364,7 +364,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
markedForEvict
-private volatileboolean markedForEvict
+private volatileboolean markedForEvict
@@ -373,7 +373,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
refCount
-privatehttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicInteger.html?is-external=true;
title="class or interface in java.util.concurrent.atomic">AtomicInteger refCount
+privatehttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicInteger.html?is-external=true;
title="class or interface in java.util.concurrent.atomic">AtomicInteger refCount
@@ -382,7 +382,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
cachedTime
-private finallong cachedTime
+private finallong cachedTime
Time this block was cached. Presumes we are created just
before we are added to the cache.
@@ -400,7 +400,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
BucketEntry
-BucketEntry(longoffset,
+BucketEntry(longoffset,
intlength,
longaccessCounter,
booleaninMemory)
@@ -420,7 +420,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
offset
-longoffset()
+longoffset()
@@ -429,7 +429,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
setOffset
-privatevoidsetOffset(longvalue)
+privatevoidsetOffset(longvalue)
@@ -438,7 +438,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
getLength
-publicintgetLength()
+publicintgetLength()
@@ -447,7 +447,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Serializabl
deserializerReference
-protectedCacheableDeserializerCacheabledeserializerReference(UniqueIndexMaphttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
title="class or interface in
java.lang">IntegerdeserialiserMap)
+protectedCacheableDeserializerCacheabledeserializerReference(UniqueIndexMaphttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true;
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d2b28a1a/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html
b/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html
index 3400507..2baa140 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/PrivateCellUtil.FirstOnRowColCell.html
@@ -28,3034 +28,2926 @@
020import static
org.apache.hadoop.hbase.HConstants.EMPTY_BYTE_ARRAY;
021import static
org.apache.hadoop.hbase.Tag.TAG_LENGTH_SIZE;
022
-023import
com.google.common.annotations.VisibleForTesting;
-024
-025import java.io.DataOutput;
-026import java.io.DataOutputStream;
-027import java.io.IOException;
-028import java.io.OutputStream;
-029import java.math.BigDecimal;
-030import java.nio.ByteBuffer;
-031import java.util.ArrayList;
-032import java.util.Iterator;
-033import java.util.List;
-034import java.util.Optional;
-035
-036import
org.apache.hadoop.hbase.KeyValue.Type;
-037import
org.apache.hadoop.hbase.filter.ByteArrayComparable;
-038import
org.apache.hadoop.hbase.io.HeapSize;
-039import
org.apache.hadoop.hbase.io.TagCompressionContext;
-040import
org.apache.hadoop.hbase.io.util.Dictionary;
-041import
org.apache.hadoop.hbase.io.util.StreamUtils;
-042import
org.apache.hadoop.hbase.util.ByteBufferUtils;
-043import
org.apache.hadoop.hbase.util.ByteRange;
-044import
org.apache.hadoop.hbase.util.Bytes;
-045import
org.apache.hadoop.hbase.util.ClassSize;
-046import
org.apache.yetus.audience.InterfaceAudience;
-047
-048
-049/**
-050 * Utility methods helpful slinging
{@link Cell} instances. It has more powerful and
-051 * rich set of APIs than those in {@link
CellUtil} for internal usage.
-052 */
-053@InterfaceAudience.Private
-054public final class PrivateCellUtil {
-055
-056 /**
-057 * Private constructor to keep this
class from being instantiated.
-058 */
-059 private PrivateCellUtil() {
-060 }
+023import java.io.DataOutput;
+024import java.io.DataOutputStream;
+025import java.io.IOException;
+026import java.io.OutputStream;
+027import java.math.BigDecimal;
+028import java.nio.ByteBuffer;
+029import java.util.ArrayList;
+030import java.util.Iterator;
+031import java.util.List;
+032import java.util.Optional;
+033import
org.apache.hadoop.hbase.KeyValue.Type;
+034import
org.apache.hadoop.hbase.filter.ByteArrayComparable;
+035import
org.apache.hadoop.hbase.io.HeapSize;
+036import
org.apache.hadoop.hbase.io.TagCompressionContext;
+037import
org.apache.hadoop.hbase.io.util.Dictionary;
+038import
org.apache.hadoop.hbase.io.util.StreamUtils;
+039import
org.apache.hadoop.hbase.util.ByteBufferUtils;
+040import
org.apache.hadoop.hbase.util.ByteRange;
+041import
org.apache.hadoop.hbase.util.Bytes;
+042import
org.apache.hadoop.hbase.util.ClassSize;
+043import
org.apache.yetus.audience.InterfaceAudience;
+044
+045import
org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
+046
+047/**
+048 * Utility methods helpful slinging
{@link Cell} instances. It has more powerful and
+049 * rich set of APIs than those in {@link
CellUtil} for internal usage.
+050 */
+051@InterfaceAudience.Private
+052public final class PrivateCellUtil {
+053
+054 /**
+055 * Private constructor to keep this
class from being instantiated.
+056 */
+057 private PrivateCellUtil() {
+058 }
+059
+060 /*** ByteRange
***/
061
-062 /*** ByteRange
***/
-063
-064 public static ByteRange
fillRowRange(Cell cell, ByteRange range) {
-065return range.set(cell.getRowArray(),
cell.getRowOffset(), cell.getRowLength());
-066 }
-067
-068 public static ByteRange
fillFamilyRange(Cell cell, ByteRange range) {
-069return
range.set(cell.getFamilyArray(), cell.getFamilyOffset(),
cell.getFamilyLength());
-070 }
-071
-072 public static ByteRange
fillQualifierRange(Cell cell, ByteRange range) {
-073return
range.set(cell.getQualifierArray(), cell.getQualifierOffset(),
-074 cell.getQualifierLength());
-075 }
-076
-077 public static ByteRange
fillValueRange(Cell cell, ByteRange range) {
-078return
range.set(cell.getValueArray(), cell.getValueOffset(),
cell.getValueLength());
-079 }
-080
-081 public static ByteRange
fillTagRange(Cell cell, ByteRange range) {
-082return range.set(cell.getTagsArray(),
cell.getTagsOffset(), cell.getTagsLength());
-083 }
+062 public static ByteRange
fillRowRange(Cell cell, ByteRange range) {
+063return range.set(cell.getRowArray(),
cell.getRowOffset(), cell.getRowLength());
+064 }
+065
+066 public static ByteRange
fillFamilyRange(Cell cell, ByteRange range) {
+067return
range.set(cell.getFamilyArray(),
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b618ac40/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
--
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
index 9d49b9a..c36fdce 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/master/HMaster.html
@@ -165,3380 +165,3375 @@
157import
org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy;
158import
org.apache.hadoop.hbase.regionserver.compactions.FIFOCompactionPolicy;
159import
org.apache.hadoop.hbase.replication.ReplicationException;
-160import
org.apache.hadoop.hbase.replication.ReplicationFactory;
-161import
org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
-162import
org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
-163import
org.apache.hadoop.hbase.replication.ReplicationQueuesZKImpl;
-164import
org.apache.hadoop.hbase.replication.master.ReplicationPeerConfigUpgrader;
-165import
org.apache.hadoop.hbase.replication.regionserver.Replication;
-166import
org.apache.hadoop.hbase.security.AccessDeniedException;
-167import
org.apache.hadoop.hbase.security.UserProvider;
-168import
org.apache.hadoop.hbase.trace.TraceUtil;
-169import
org.apache.hadoop.hbase.util.Addressing;
-170import
org.apache.hadoop.hbase.util.Bytes;
-171import
org.apache.hadoop.hbase.util.CompressionTest;
-172import
org.apache.hadoop.hbase.util.EncryptionTest;
-173import
org.apache.hadoop.hbase.util.FSUtils;
-174import
org.apache.hadoop.hbase.util.HFileArchiveUtil;
-175import
org.apache.hadoop.hbase.util.HasThread;
-176import
org.apache.hadoop.hbase.util.IdLock;
-177import
org.apache.hadoop.hbase.util.ModifyRegionUtils;
-178import
org.apache.hadoop.hbase.util.Pair;
-179import
org.apache.hadoop.hbase.util.Threads;
-180import
org.apache.hadoop.hbase.util.VersionInfo;
-181import
org.apache.hadoop.hbase.util.ZKDataMigrator;
-182import
org.apache.hadoop.hbase.zookeeper.LoadBalancerTracker;
-183import
org.apache.hadoop.hbase.zookeeper.MasterAddressTracker;
-184import
org.apache.hadoop.hbase.zookeeper.MasterMaintenanceModeTracker;
-185import
org.apache.hadoop.hbase.zookeeper.RegionNormalizerTracker;
-186import
org.apache.hadoop.hbase.zookeeper.ZKClusterId;
-187import
org.apache.hadoop.hbase.zookeeper.ZKUtil;
-188import
org.apache.hadoop.hbase.zookeeper.ZKWatcher;
-189import
org.apache.hadoop.hbase.zookeeper.ZNodePaths;
-190import
org.apache.yetus.audience.InterfaceAudience;
-191import
org.apache.zookeeper.KeeperException;
-192import org.eclipse.jetty.server.Server;
-193import
org.eclipse.jetty.server.ServerConnector;
-194import
org.eclipse.jetty.servlet.ServletHolder;
-195import
org.eclipse.jetty.webapp.WebAppContext;
-196import org.slf4j.Logger;
-197import org.slf4j.LoggerFactory;
-198
-199import
org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
-200import
org.apache.hadoop.hbase.shaded.com.google.common.collect.Lists;
-201import
org.apache.hadoop.hbase.shaded.com.google.common.collect.Maps;
-202import
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-203import
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState;
-204import
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionServerInfo;
-205import
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas;
-206import
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceViolationPolicy;
-207import
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription;
-208import
org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos;
-209
-210/**
-211 * HMaster is the "master server" for
HBase. An HBase cluster has one active
-212 * master. If many masters are started,
all compete. Whichever wins goes on to
-213 * run the cluster. All others park
themselves in their constructor until
-214 * master or cluster shutdown or until
the active master loses its lease in
-215 * zookeeper. Thereafter, all running
master jostle to take over master role.
-216 *
-217 * pThe Master can be asked
shutdown the cluster. See {@link #shutdown()}. In
-218 * this case it will tell all
regionservers to go down and then wait on them
-219 * all reporting in that they are down.
This master will then shut itself down.
+160import
org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
+161import
org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
+162import
org.apache.hadoop.hbase.replication.master.ReplicationPeerConfigUpgrader;
+163import
org.apache.hadoop.hbase.replication.regionserver.Replication;
+164import
org.apache.hadoop.hbase.security.AccessDeniedException;
+165import
org.apache.hadoop.hbase.security.UserProvider;
+166import
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7c0589c0/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TruncateTableFuture.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TruncateTableFuture.html
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TruncateTableFuture.html
index 6fecbc9..2accda0 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TruncateTableFuture.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.TruncateTableFuture.html
@@ -34,4140 +34,4141 @@
026import
java.nio.charset.StandardCharsets;
027import java.util.ArrayList;
028import java.util.Arrays;
-029import java.util.Collection;
-030import java.util.EnumSet;
-031import java.util.HashMap;
-032import java.util.Iterator;
-033import java.util.LinkedList;
-034import java.util.List;
-035import java.util.Map;
-036import java.util.Set;
-037import java.util.concurrent.Callable;
-038import
java.util.concurrent.ExecutionException;
-039import java.util.concurrent.Future;
-040import java.util.concurrent.TimeUnit;
-041import
java.util.concurrent.TimeoutException;
-042import
java.util.concurrent.atomic.AtomicInteger;
-043import
java.util.concurrent.atomic.AtomicReference;
-044import java.util.regex.Pattern;
-045import java.util.stream.Collectors;
-046import java.util.stream.Stream;
-047import
org.apache.hadoop.conf.Configuration;
-048import
org.apache.hadoop.hbase.Abortable;
-049import
org.apache.hadoop.hbase.CacheEvictionStats;
-050import
org.apache.hadoop.hbase.CacheEvictionStatsBuilder;
-051import
org.apache.hadoop.hbase.ClusterMetrics.Option;
-052import
org.apache.hadoop.hbase.ClusterStatus;
-053import
org.apache.hadoop.hbase.DoNotRetryIOException;
-054import
org.apache.hadoop.hbase.HBaseConfiguration;
-055import
org.apache.hadoop.hbase.HConstants;
-056import
org.apache.hadoop.hbase.HRegionInfo;
-057import
org.apache.hadoop.hbase.HRegionLocation;
-058import
org.apache.hadoop.hbase.HTableDescriptor;
-059import
org.apache.hadoop.hbase.MasterNotRunningException;
-060import
org.apache.hadoop.hbase.MetaTableAccessor;
-061import
org.apache.hadoop.hbase.NamespaceDescriptor;
-062import
org.apache.hadoop.hbase.NamespaceNotFoundException;
-063import
org.apache.hadoop.hbase.NotServingRegionException;
-064import
org.apache.hadoop.hbase.RegionLoad;
-065import
org.apache.hadoop.hbase.RegionLocations;
-066import
org.apache.hadoop.hbase.ServerName;
-067import
org.apache.hadoop.hbase.TableExistsException;
-068import
org.apache.hadoop.hbase.TableName;
-069import
org.apache.hadoop.hbase.TableNotDisabledException;
-070import
org.apache.hadoop.hbase.TableNotFoundException;
-071import
org.apache.hadoop.hbase.UnknownRegionException;
-072import
org.apache.hadoop.hbase.ZooKeeperConnectionException;
-073import
org.apache.hadoop.hbase.client.replication.ReplicationPeerConfigUtil;
-074import
org.apache.hadoop.hbase.client.replication.TableCFs;
-075import
org.apache.hadoop.hbase.client.security.SecurityCapability;
-076import
org.apache.hadoop.hbase.exceptions.TimeoutIOException;
-077import
org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
-078import
org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils;
-079import
org.apache.hadoop.hbase.ipc.HBaseRpcController;
-080import
org.apache.hadoop.hbase.ipc.RpcControllerFactory;
-081import
org.apache.hadoop.hbase.quotas.QuotaFilter;
-082import
org.apache.hadoop.hbase.quotas.QuotaRetriever;
-083import
org.apache.hadoop.hbase.quotas.QuotaSettings;
-084import
org.apache.hadoop.hbase.regionserver.wal.FailedLogCloseException;
-085import
org.apache.hadoop.hbase.replication.ReplicationException;
-086import
org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
-087import
org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
-088import
org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils;
-089import
org.apache.hadoop.hbase.snapshot.HBaseSnapshotException;
-090import
org.apache.hadoop.hbase.snapshot.RestoreSnapshotException;
-091import
org.apache.hadoop.hbase.snapshot.SnapshotCreationException;
-092import
org.apache.hadoop.hbase.snapshot.UnknownSnapshotException;
-093import
org.apache.hadoop.hbase.util.Addressing;
-094import
org.apache.hadoop.hbase.util.Bytes;
-095import
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-096import
org.apache.hadoop.hbase.util.ForeignExceptionUtil;
-097import
org.apache.hadoop.hbase.util.Pair;
-098import
org.apache.hadoop.ipc.RemoteException;
-099import
org.apache.hadoop.util.StringUtils;
-100import
org.apache.yetus.audience.InterfaceAudience;
-101import
org.apache.yetus.audience.InterfaceStability;
-102import org.slf4j.Logger;
-103import org.slf4j.LoggerFactory;
-104
-105import
org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
-106import
org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4b2cc17/devapidocs/org/apache/hadoop/hbase/ClusterStatus.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/ClusterStatus.html
b/devapidocs/org/apache/hadoop/hbase/ClusterStatus.html
index f60f387..2d6cdec 100644
--- a/devapidocs/org/apache/hadoop/hbase/ClusterStatus.html
+++ b/devapidocs/org/apache/hadoop/hbase/ClusterStatus.html
@@ -18,8 +18,8 @@
catch(err) {
}
//-->
-var methods =
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":42,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":42,"i20":10,"i21":10,"i22":42,"i23":10,"i24":10,"i25":9,"i26":10};
-var tabs = {65535:["t0","All Methods"],1:["t1","Static
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete
Methods"],32:["t6","Deprecated Methods"]};
+var methods =
{"i0":42,"i1":42,"i2":42,"i3":42,"i4":42,"i5":42,"i6":42,"i7":42,"i8":42,"i9":42,"i10":42,"i11":42,"i12":42,"i13":42,"i14":42,"i15":42,"i16":42,"i17":42,"i18":42,"i19":42,"i20":42,"i21":42,"i22":42,"i23":42,"i24":42,"i25":42,"i26":42,"i27":42,"i28":42};
+var tabs = {65535:["t0","All Methods"],2:["t2","Instance
Methods"],8:["t4","Concrete Methods"],32:["t6","Deprecated Methods"]};
var altColor = "altColor";
var rowColor = "rowColor";
var tableTab = "tableTab";
@@ -49,8 +49,8 @@ var activeTableTab = "activeTableTab";
-PrevClass
-NextClass
+PrevClass
+NextClass
Frames
@@ -74,7 +74,7 @@ var activeTableTab = "activeTableTab";
Summary:
-Nested|
+Nested|
Field|
Constr|
Method
@@ -107,11 +107,21 @@ var activeTableTab = "activeTableTab";
+
+All Implemented Interfaces:
+ClusterMetrics
+
+Deprecated.
+As of release 2.0.0, this
will be removed in HBase 3.0.0
+ Use ClusterMetrics
instead.
+
@InterfaceAudience.Public
-public class ClusterStatus
-extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
+ http://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true;
title="class or interface in java.lang">@Deprecated
+public class ClusterStatus
+extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
+implements ClusterMetrics
Status information on the HBase cluster.
ClusterStatus provides clients with information such as:
@@ -128,7 +138,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
Regions in transition at master
The unique cluster ID
- ClusterStatus.Option provides a
way to get desired ClusterStatus information.
+ ClusterMetrics.Option provides a
way to get desired ClusterStatus information.
The following codes will get all the cluster information.
@@ -159,25 +169,13 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
Nested Class Summary
-
-Nested Classes
-
-Modifier and Type
-Class and Description
-
-
-static class
-ClusterStatus.Builder
-Builder for construct a ClusterStatus.
-
-
-
-static class
-ClusterStatus.Option
-Kinds of ClusterStatus
-
-
-
+
+
+
+
+Nested classes/interfaces inherited from
interfaceorg.apache.hadoop.hbase.ClusterMetrics
+ClusterMetrics.Option
+
@@ -193,48 +191,16 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
Field and Description
-private http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListServerName
-backupMasters
+private ClusterMetrics
+metrics
+Deprecated.
+
-private http://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true;
title="class or interface in java.lang">Boolean
-balancerOn
-
-
-private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
-clusterId
-
-
-private http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListServerName
-deadServers
-
-
-private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String
-hbaseVersion
-
-
-private http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true;
title="class or interface in java.util">ListRegionState
-intransition
-
-
-private http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true;
title="class or interface in java.util">MapServerName,ServerLoad
-liveServers
-
-
-private ServerName
-master
-
-
-private http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String[]
-masterCoprocessors
-
-
-private int
-masterInfoPort
-
-
private static byte
-VERSION
+VERSION
+Deprecated.
+
@@ -248,12 +214,15 @@ extends
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/505bbb2e/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
b/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
index 129390f..a21c2b6 100644
--- a/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
+++ b/devapidocs/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
@@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab";
@InterfaceAudience.Private
-public final class BackupSystemTable
+public final class BackupSystemTable
extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Object
implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true;
title="class or interface in java.io">Closeable
This class provides API to access backup system table
@@ -234,7 +234,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
INCR_BACKUP_SET
-private static
org.apache.commons.logging.Log
+private static org.slf4j.Logger
LOG
@@ -953,7 +953,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
LOG
-private static finalorg.apache.commons.logging.Log LOG
+private static finalorg.slf4j.Logger LOG
@@ -962,7 +962,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
tableName
-privateTableName tableName
+privateTableName tableName
@@ -971,7 +971,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
SESSIONS_FAMILY
-static finalbyte[] SESSIONS_FAMILY
+static finalbyte[] SESSIONS_FAMILY
Stores backup sessions (contexts)
@@ -981,7 +981,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
META_FAMILY
-static finalbyte[] META_FAMILY
+static finalbyte[] META_FAMILY
Stores other meta
@@ -991,7 +991,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
BULK_LOAD_FAMILY
-static finalbyte[] BULK_LOAD_FAMILY
+static finalbyte[] BULK_LOAD_FAMILY
@@ -1000,7 +1000,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
connection
-private finalConnection connection
+private finalConnection connection
Connection to HBase cluster, shared among all
instances
@@ -1010,7 +1010,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
BACKUP_INFO_PREFIX
-private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String BACKUP_INFO_PREFIX
+private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String BACKUP_INFO_PREFIX
See Also:
Constant
Field Values
@@ -1023,7 +1023,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
START_CODE_ROW
-private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String START_CODE_ROW
+private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String START_CODE_ROW
See Also:
Constant
Field Values
@@ -1036,7 +1036,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
ACTIVE_SESSION_ROW
-private static finalbyte[] ACTIVE_SESSION_ROW
+private static finalbyte[] ACTIVE_SESSION_ROW
@@ -1045,7 +1045,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
ACTIVE_SESSION_COL
-private static finalbyte[] ACTIVE_SESSION_COL
+private static finalbyte[] ACTIVE_SESSION_COL
@@ -1054,7 +1054,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
ACTIVE_SESSION_YES
-private static finalbyte[] ACTIVE_SESSION_YES
+private static finalbyte[] ACTIVE_SESSION_YES
@@ -1063,7 +1063,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
ACTIVE_SESSION_NO
-private static finalbyte[] ACTIVE_SESSION_NO
+private static finalbyte[] ACTIVE_SESSION_NO
@@ -1072,7 +1072,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
INCR_BACKUP_SET
-private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String INCR_BACKUP_SET
+private static finalhttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true;
title="class or interface in java.lang">String INCR_BACKUP_SET
See Also:
Constant
Field Values
@@ -1085,7 +1085,7 @@ implements http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.h
TABLE_RS_LOG_MAP_PREFIX
-private static
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dad9a249/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
index 067d702..d42cfb1 100644
---
a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
+++
b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.IOErrorWithCause.html
@@ -137,7 +137,7 @@ var activeTableTab = "activeTableTab";
-public static class ThriftServerRunner.IOErrorWithCause
+public static class ThriftServerRunner.IOErrorWithCause
extends org.apache.hadoop.hbase.thrift.generated.IOError
See Also:
@@ -273,7 +273,7 @@ extends
org.apache.hadoop.hbase.thrift.generated.IOError
cause
-privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
title="class or interface in java.lang">Throwable cause
+privatehttp://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
title="class or interface in java.lang">Throwable cause
@@ -290,7 +290,7 @@ extends
org.apache.hadoop.hbase.thrift.generated.IOError
IOErrorWithCause
-publicIOErrorWithCause(http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
title="class or interface in java.lang">Throwablecause)
+publicIOErrorWithCause(http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
title="class or interface in java.lang">Throwablecause)
@@ -307,7 +307,7 @@ extends
org.apache.hadoop.hbase.thrift.generated.IOError
getCause
-publichttp://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
title="class or interface in java.lang">ThrowablegetCause()
+publichttp://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
title="class or interface in java.lang">ThrowablegetCause()
Overrides:
http://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true#getCause--;
title="class or interface in java.lang">getCausein
classhttp://docs.oracle.com/javase/8/docs/api/java/lang/Throwable.html?is-external=true;
title="class or interface in java.lang">Throwable
@@ -320,7 +320,7 @@ extends
org.apache.hadoop.hbase.thrift.generated.IOError
equals
-publicbooleanequals(http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Objectother)
+publicbooleanequals(http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true;
title="class or interface in java.lang">Objectother)
Overrides:
equalsin
classorg.apache.hadoop.hbase.thrift.generated.IOError
@@ -333,7 +333,7 @@ extends
org.apache.hadoop.hbase.thrift.generated.IOError
hashCode
-publicinthashCode()
+publicinthashCode()
Overrides:
hashCodein
classorg.apache.hadoop.hbase.thrift.generated.IOError
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dad9a249/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
index cd86385..b8aeaa7 100644
--- a/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
+++ b/devapidocs/org/apache/hadoop/hbase/thrift/ThriftServerRunner.ImplType.html
@@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
-static enum ThriftServerRunner.ImplType
+static enum ThriftServerRunner.ImplType
extends http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true;
title="class or interface in java.lang">EnumThriftServerRunner.ImplType
An enum of server implementation selections
@@ -281,7 +281,7 @@ the order they are declared.
HS_HA
-public static finalThriftServerRunner.ImplType
HS_HA
+public static finalThriftServerRunner.ImplType
HS_HA
@@ -290,7 +290,7 @@ the order they are declared.
NONBLOCKING
-public static finalThriftServerRunner.ImplType
NONBLOCKING
+public static finalThriftServerRunner.ImplType
NONBLOCKING
@@ -299,7 +299,7 @@ the order they are declared.
THREAD_POOL
-public static finalThriftServerRunner.ImplType
THREAD_POOL
+public static finalThriftServerRunner.ImplType
THREAD_POOL
@@ -308,7 +308,7 @@ the order they are declared.
THREADED_SELECTOR
-public static finalThriftServerRunner.ImplType
THREADED_SELECTOR
+public static finalThriftServerRunner.ImplType
THREADED_SELECTOR
@@ -325,7 +325,7 @@ the order they are declared.
DEFAULT
-public static finalThriftServerRunner.ImplType
DEFAULT
+public static finalThriftServerRunner.ImplType
DEFAULT
@@ -334,7 +334,7 @@ the order
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c4c0cfa5/apidocs/src-html/org/apache/hadoop/hbase/util/Base64.Base64OutputStream.html
--
diff --git
a/apidocs/src-html/org/apache/hadoop/hbase/util/Base64.Base64OutputStream.html
b/apidocs/src-html/org/apache/hadoop/hbase/util/Base64.Base64OutputStream.html
index 6646ba1..70481ce 100644
---
a/apidocs/src-html/org/apache/hadoop/hbase/util/Base64.Base64OutputStream.html
+++
b/apidocs/src-html/org/apache/hadoop/hbase/util/Base64.Base64OutputStream.html
@@ -43,1651 +43,1652 @@
035import java.io.OutputStream;
036import java.io.Serializable;
037import
java.io.UnsupportedEncodingException;
-038import java.util.zip.GZIPInputStream;
-039import java.util.zip.GZIPOutputStream;
-040
-041import org.apache.commons.logging.Log;
-042import
org.apache.commons.logging.LogFactory;
-043import
org.apache.yetus.audience.InterfaceAudience;
-044
-045/**
-046 * Encodes and decodes to and from Base64
notation.
-047 *
-048 * p
-049 * Homepage: a
href="http://iharder.net/base64"http://iharder.net/base64/a;.
-050 * /p
-051 *
-052 * p
-053 * Change Log:
-054 * /p
-055 * ul
-056 * liv2.2.1 - Fixed bug using
URL_SAFE and ORDERED encodings. Fixed bug
-057 * when using very small files
(~lt; 40 bytes)./li
-058 * liv2.2 - Added some helper
methods for encoding/decoding directly from
-059 * one file to the next. Also added a
main() method to support command
-060 * line encoding/decoding from one
file to the next. Also added these
-061 * Base64 dialects:
-062 * ol
-063 * liThe default is RFC3548
format./li
-064 * liUsing Base64.URLSAFE
generates URL and file name friendly format as
-065 * described in Section 4 of
RFC3548.
-066 *
http://www.faqs.org/rfcs/rfc3548.html/li;
-067 * liUsing Base64.ORDERED
generates URL and file name friendly format
-068 * that preserves lexical
ordering as described in
-069 *
http://www.faqs.org/qa/rfcc-1940.html/li;
-070 * /ol
-071 * p
-072 * Special thanks to Jim Kellerman at
a href="http://www.powerset.com/";
-073 * http://www.powerset.com//a;
for contributing the new Base64 dialects.
-074 * /li
-075 *
-076 * liv2.1 - Cleaned up javadoc
comments and unused variables and methods.
-077 * Added some convenience methods for
reading and writing to and from files.
-078 * /li
-079 * liv2.0.2 - Now specifies
UTF-8 encoding in places where the code fails on
-080 * systems with other encodings (like
EBCDIC)./li
-081 * liv2.0.1 - Fixed an error
when decoding a single byte, that is, when the
-082 * encoded data was a single
byte./li
-083 * liv2.0 - I got rid of
methods that used booleans to set options. Now
-084 * everything is more consolidated
and cleaner. The code now detects when
-085 * data that's being decoded is
gzip-compressed and will decompress it
-086 * automatically. Generally things
are cleaner. You'll probably have to
-087 * change some method calls that you
were making to support the new options
-088 * format (ttint/tts
that you "OR" together)./li
-089 * liv1.5.1 - Fixed bug when
decompressing and decoding to a byte[] using
-090 * ttdecode( String s,
boolean gzipCompressed )/tt. Added the ability to
-091 * "suspend" encoding in the Output
Stream so you can turn on and off the
-092 * encoding if you need to embed
base64 data in an otherwise "normal" stream
-093 * (like an XML file)./li
-094 * liv1.5 - Output stream pases
on flush() command but doesn't do anything
-095 * itself. This helps when using GZIP
streams. Added the ability to
-096 * GZip-compress objects before
encoding them./li
-097 * liv1.4 - Added helper
methods to read/write files./li
-098 * liv1.3.6 - Fixed
OutputStream.flush() so that 'position' is reset./li
-099 * liv1.3.5 - Added flag to
turn on and off line breaks. Fixed bug in input
-100 * stream where last buffer being
read, if not completely full, was not
-101 * returned./li
-102 * liv1.3.4 - Fixed when
"improperly padded stream" error was thrown at the
-103 * wrong time./li
-104 * liv1.3.3 - Fixed I/O streams
which were totally messed up./li
-105 * /ul
-106 *
-107 * p
-108 * I am placing this code in the Public
Domain. Do with it as you will. This
-109 * software comes with no guarantees or
warranties but with plenty of
-110 * well-wishing instead!
-111 * p
-112 * Please visit a
href="http://iharder.net/base64"http://iharder.net/base64/a;
-113 * periodically to check for updates or
to contribute improvements.
-114 * p
-115 * author: Robert Harder,
r...@iharder.net
-116 * br
-117 * version: 2.2.1
-118 */
-119@InterfaceAudience.Public
-120public class Base64 {
-121
-122 /* P U B L I C F I E L D S
*/
-123
-124 /** No options specified. Value is
zero. */
-125 public final static int NO_OPTIONS =
0;
-126
-127
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/071f974b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html
index 5e45072..34535d8 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/ipc/RpcServer.BlockingServiceAndInterface.html
@@ -50,754 +50,751 @@
042import
org.apache.hadoop.hbase.CallQueueTooBigException;
043import
org.apache.hadoop.hbase.CellScanner;
044import
org.apache.hadoop.hbase.DoNotRetryIOException;
-045import
org.apache.hadoop.hbase.HBaseInterfaceAudience;
-046import
org.apache.hadoop.hbase.HConstants;
-047import org.apache.hadoop.hbase.Server;
-048import
org.apache.hadoop.hbase.conf.ConfigurationObserver;
-049import
org.apache.hadoop.hbase.exceptions.RequestTooBigException;
-050import
org.apache.hadoop.hbase.io.ByteBufferPool;
-051import
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
-052import
org.apache.hadoop.hbase.monitoring.TaskMonitor;
-053import
org.apache.hadoop.hbase.nio.ByteBuff;
-054import
org.apache.hadoop.hbase.nio.MultiByteBuff;
-055import
org.apache.hadoop.hbase.nio.SingleByteBuff;
-056import
org.apache.hadoop.hbase.regionserver.RSRpcServices;
-057import
org.apache.hadoop.hbase.security.SaslUtil;
-058import
org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection;
-059import
org.apache.hadoop.hbase.security.User;
-060import
org.apache.hadoop.hbase.security.UserProvider;
-061import
org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager;
-062import
org.apache.hadoop.hbase.util.Pair;
-063import
org.apache.hadoop.security.UserGroupInformation;
-064import
org.apache.hadoop.security.authorize.AuthorizationException;
-065import
org.apache.hadoop.security.authorize.PolicyProvider;
-066import
org.apache.hadoop.security.authorize.ServiceAuthorizationManager;
-067import
org.apache.hadoop.security.token.SecretManager;
-068import
org.apache.hadoop.security.token.TokenIdentifier;
-069import
org.apache.yetus.audience.InterfaceAudience;
-070import
org.apache.yetus.audience.InterfaceStability;
-071
-072import
org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
-073import
org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService;
-074import
org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor;
-075import
org.apache.hadoop.hbase.shaded.com.google.protobuf.Message;
-076import
org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
-077import
org.apache.hadoop.hbase.shaded.com.google.protobuf.TextFormat;
-078import
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-079import
org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos;
-080import
org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader;
-081
-082/**
-083 * An RPC server that hosts protobuf
described Services.
-084 *
-085 */
-086@InterfaceAudience.LimitedPrivate({HBaseInterfaceAudience.COPROC,
HBaseInterfaceAudience.PHOENIX})
-087@InterfaceStability.Evolving
-088public abstract class RpcServer
implements RpcServerInterface,
-089ConfigurationObserver {
-090 // LOG is being used in CallRunner and
the log level is being changed in tests
-091 public static final Log LOG =
LogFactory.getLog(RpcServer.class);
-092 protected static final
CallQueueTooBigException CALL_QUEUE_TOO_BIG_EXCEPTION
-093 = new CallQueueTooBigException();
+045import
org.apache.hadoop.hbase.HConstants;
+046import org.apache.hadoop.hbase.Server;
+047import
org.apache.hadoop.hbase.conf.ConfigurationObserver;
+048import
org.apache.hadoop.hbase.exceptions.RequestTooBigException;
+049import
org.apache.hadoop.hbase.io.ByteBufferPool;
+050import
org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler;
+051import
org.apache.hadoop.hbase.monitoring.TaskMonitor;
+052import
org.apache.hadoop.hbase.nio.ByteBuff;
+053import
org.apache.hadoop.hbase.nio.MultiByteBuff;
+054import
org.apache.hadoop.hbase.nio.SingleByteBuff;
+055import
org.apache.hadoop.hbase.regionserver.RSRpcServices;
+056import
org.apache.hadoop.hbase.security.SaslUtil;
+057import
org.apache.hadoop.hbase.security.SaslUtil.QualityOfProtection;
+058import
org.apache.hadoop.hbase.security.User;
+059import
org.apache.hadoop.hbase.security.UserProvider;
+060import
org.apache.hadoop.hbase.security.token.AuthenticationTokenSecretManager;
+061import
org.apache.hadoop.hbase.util.Pair;
+062import
org.apache.hadoop.security.UserGroupInformation;
+063import
org.apache.hadoop.security.authorize.AuthorizationException;
+064import
org.apache.hadoop.security.authorize.PolicyProvider;
+065import
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/dc4e5c85/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutput.State.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutput.State.html
b/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutput.State.html
index dd54dd2..a9c1142 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutput.State.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutput.State.html
@@ -388,195 +388,196 @@
380
381 @Override
382 public DatanodeInfo[] getPipeline() {
-383return locations;
-384 }
-385
-386 private void
flushBuffer(CompletableFutureLong future, ByteBuf dataBuf,
-387 long nextPacketOffsetInBlock,
boolean syncBlock) {
-388int dataLen =
dataBuf.readableBytes();
-389int chunkLen =
summer.getBytesPerChecksum();
-390int trailingPartialChunkLen = dataLen
% chunkLen;
-391int numChecks = dataLen / chunkLen +
(trailingPartialChunkLen != 0 ? 1 : 0);
-392int checksumLen = numChecks *
summer.getChecksumSize();
-393ByteBuf checksumBuf =
alloc.directBuffer(checksumLen);
-394
summer.calculateChunkedSums(dataBuf.nioBuffer(), checksumBuf.nioBuffer(0,
checksumLen));
-395
checksumBuf.writerIndex(checksumLen);
-396PacketHeader header = new
PacketHeader(4 + checksumLen + dataLen, nextPacketOffsetInBlock,
-397nextPacketSeqno, false, dataLen,
syncBlock);
-398int headerLen =
header.getSerializedSize();
-399ByteBuf headerBuf =
alloc.buffer(headerLen);
-400
header.putInBuffer(headerBuf.nioBuffer(0, headerLen));
-401headerBuf.writerIndex(headerLen);
-402Callback c = new Callback(future,
nextPacketOffsetInBlock + dataLen, datanodeList);
-403waitingAckQueue.addLast(c);
-404// recheck again after we pushed the
callback to queue
-405if (state != State.STREAMING
waitingAckQueue.peekFirst() == c) {
-406 future.completeExceptionally(new
IOException("stream already broken"));
-407 // it's the one we have just pushed
or just a no-op
-408 waitingAckQueue.removeFirst();
-409 return;
-410}
-411datanodeList.forEach(ch - {
-412
ch.write(headerBuf.retainedDuplicate());
-413
ch.write(checksumBuf.retainedDuplicate());
-414
ch.writeAndFlush(dataBuf.retainedDuplicate());
-415});
-416checksumBuf.release();
-417headerBuf.release();
-418dataBuf.release();
-419nextPacketSeqno++;
-420 }
-421
-422 private void
flush0(CompletableFutureLong future, boolean syncBlock) {
-423if (state != State.STREAMING) {
-424 future.completeExceptionally(new
IOException("stream already broken"));
-425 return;
-426}
-427int dataLen = buf.readableBytes();
-428if (dataLen ==
trailingPartialChunkLength) {
-429 // no new data
-430 long lengthAfterFlush =
nextPacketOffsetInBlock + dataLen;
-431 Callback lastFlush =
waitingAckQueue.peekLast();
-432 if (lastFlush != null) {
-433Callback c = new Callback(future,
lengthAfterFlush, Collections.emptyList());
-434waitingAckQueue.addLast(c);
-435// recheck here if we have
already removed the previous callback from the queue
-436if (waitingAckQueue.peekFirst()
== c) {
-437 // all previous callbacks have
been removed
-438 // notice that this does mean
we will always win here because the background thread may
-439 // have already started to mark
the future here as completed in the completed or failed
-440 // methods but haven't removed
it from the queue yet. That's also why the removeFirst
-441 // call below may be a no-op.
-442 if (state != State.STREAMING)
{
-443
future.completeExceptionally(new IOException("stream already broken"));
-444 } else {
-445
future.complete(lengthAfterFlush);
-446 }
-447 // it's the one we have just
pushed or just a no-op
-448
waitingAckQueue.removeFirst();
-449}
-450 } else {
-451// we must have acked all the
data so the ackedBlockLength must be same with
-452// lengthAfterFlush
-453
future.complete(lengthAfterFlush);
-454 }
-455 return;
-456}
-457
-458if (encryptor != null) {
-459 ByteBuf encryptBuf =
alloc.directBuffer(dataLen);
-460 buf.readBytes(encryptBuf,
trailingPartialChunkLength);
-461 int toEncryptLength = dataLen -
trailingPartialChunkLength;
-462 try {
-463
encryptor.encrypt(buf.nioBuffer(trailingPartialChunkLength, toEncryptLength),
-464
encryptBuf.nioBuffer(trailingPartialChunkLength, toEncryptLength));
-465 } catch (IOException e) {
-466encryptBuf.release();
-467
future.completeExceptionally(e);
-468
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/4abd958d/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
index ae9896e..cb2d47f 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.html
@@ -41,25 +41,25 @@
033import java.util.TreeMap;
034import java.util.TreeSet;
035
-036import
org.apache.commons.lang3.StringUtils;
-037import org.apache.commons.logging.Log;
-038import
org.apache.commons.logging.LogFactory;
-039import
org.apache.hadoop.conf.Configuration;
-040import org.apache.hadoop.fs.Path;
-041import org.apache.hadoop.hbase.Cell;
-042import
org.apache.hadoop.hbase.CellUtil;
-043import
org.apache.hadoop.hbase.HBaseConfiguration;
-044import
org.apache.hadoop.hbase.HColumnDescriptor;
-045import
org.apache.hadoop.hbase.HTableDescriptor;
-046import
org.apache.hadoop.hbase.NamespaceDescriptor;
-047import
org.apache.hadoop.hbase.ServerName;
-048import
org.apache.hadoop.hbase.TableName;
-049import
org.apache.hadoop.hbase.backup.BackupInfo;
-050import
org.apache.hadoop.hbase.backup.BackupInfo.BackupState;
-051import
org.apache.hadoop.hbase.backup.BackupRestoreConstants;
-052import
org.apache.hadoop.hbase.backup.BackupType;
-053import
org.apache.hadoop.hbase.backup.util.BackupUtils;
-054import
org.apache.yetus.audience.InterfaceAudience;
+036import
org.apache.commons.lang3.ArrayUtils;
+037import
org.apache.commons.lang3.StringUtils;
+038import org.apache.commons.logging.Log;
+039import
org.apache.commons.logging.LogFactory;
+040import
org.apache.hadoop.conf.Configuration;
+041import org.apache.hadoop.fs.Path;
+042import org.apache.hadoop.hbase.Cell;
+043import
org.apache.hadoop.hbase.CellUtil;
+044import
org.apache.hadoop.hbase.HBaseConfiguration;
+045import
org.apache.hadoop.hbase.HColumnDescriptor;
+046import
org.apache.hadoop.hbase.HTableDescriptor;
+047import
org.apache.hadoop.hbase.NamespaceDescriptor;
+048import
org.apache.hadoop.hbase.ServerName;
+049import
org.apache.hadoop.hbase.TableName;
+050import
org.apache.hadoop.hbase.backup.BackupInfo;
+051import
org.apache.hadoop.hbase.backup.BackupInfo.BackupState;
+052import
org.apache.hadoop.hbase.backup.BackupRestoreConstants;
+053import
org.apache.hadoop.hbase.backup.BackupType;
+054import
org.apache.hadoop.hbase.backup.util.BackupUtils;
055import
org.apache.hadoop.hbase.client.Admin;
056import
org.apache.hadoop.hbase.client.Connection;
057import
org.apache.hadoop.hbase.client.Delete;
@@ -70,12 +70,12 @@
062import
org.apache.hadoop.hbase.client.Scan;
063import
org.apache.hadoop.hbase.client.SnapshotDescription;
064import
org.apache.hadoop.hbase.client.Table;
-065import
org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos;
-066import
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
-067import
org.apache.hadoop.hbase.util.ArrayUtils;
-068import
org.apache.hadoop.hbase.util.Bytes;
-069import
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
-070import
org.apache.hadoop.hbase.util.Pair;
+065import
org.apache.hadoop.hbase.util.Bytes;
+066import
org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
+067import
org.apache.hadoop.hbase.util.Pair;
+068import
org.apache.yetus.audience.InterfaceAudience;
+069import
org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos;
+070import
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
071
072/**
073 * This class provides API to access
backup system tablebr
@@ -585,1396 +585,1397 @@
577try (Table table =
connection.getTable(tableName)) {
578 Put put =
createPutForStartBackupSession();
579 // First try to put if row does not
exist
-580 if
(!table.checkAndPut(ACTIVE_SESSION_ROW, SESSIONS_FAMILY, ACTIVE_SESSION_COL,
null, put)) {
-581// Row exists, try to put if
value == ACTIVE_SESSION_NO
-582if
(!table.checkAndPut(ACTIVE_SESSION_ROW, SESSIONS_FAMILY, ACTIVE_SESSION_COL,
-583 ACTIVE_SESSION_NO, put)) {
-584 throw new IOException("There is
an active backup exclusive operation");
-585}
-586 }
-587}
-588 }
-589
-590 private Put
createPutForStartBackupSession() {
-591Put put = new
Put(ACTIVE_SESSION_ROW);
-592put.addColumn(SESSIONS_FAMILY,
ACTIVE_SESSION_COL, ACTIVE_SESSION_YES);
-593return put;
-594 }
-595
-596 public void
finishBackupExclusiveOperation() throws IOException {
-597LOG.debug("Finish backup exclusive
operation");
-598
-599try (Table table =
connection.getTable(tableName)) {
-600 Put put =
createPutForStopBackupSession();
-601 if
(!table.checkAndPut(ACTIVE_SESSION_ROW, SESSIONS_FAMILY,
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e23b49ba/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessor.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessor.html
b/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessor.html
index 2ec4ab5..21b91cb 100644
--- a/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessor.html
+++ b/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessor.html
@@ -49,7 +49,7 @@ var activeTableTab = "activeTableTab";
-PrevClass
+PrevClass
NextClass
@@ -250,7 +250,7 @@ extends
-PrevClass
+PrevClass
NextClass
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/e23b49ba/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.html
--
diff --git
a/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.html
b/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.html
index 8964468..d941a19 100644
---
a/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.html
+++
b/devapidocs/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.html
@@ -134,7 +134,7 @@ extends
Connection
createConnection(org.apache.hadoop.conf.Configurationconf)
-Creates a cluster connection using the passed
configuration.
+Creates a cluster connection using the passed
Configuration.
@@ -146,7 +146,7 @@ extends
Connection
getConnection()
-Be careful RPC'ing from a Coprocessor context.
+Returns the hosts' Connection to the Cluster.
@@ -268,22 +268,33 @@ extends
getConnection
-ConnectiongetConnection()
-Be careful RPC'ing from a Coprocessor context.
- RPC's will fail, stall, retry, and/or crawl because the remote side is not
online, is
- struggling or it is on the other side of a network partition. Any use of
Connection from
- inside a Coprocessor must be able to handle all such hiccups.
+ConnectiongetConnection()
+Returns the hosts' Connection to the Cluster. Do not
close! This is a shared connection
+ with the hosting server. Throws http://docs.oracle.com/javase/8/docs/api/java/lang/UnsupportedOperationException.html?is-external=true;
title="class or interface in
java.lang">UnsupportedOperationException if you try to close
+ or abort it.
- Using a Connection to get at a local resource -- say a Region that is on
the local
- Server or using Admin Interface from a Coprocessor hosted on the Master --
will result in a
- short-circuit of the RPC framework to make a direct invocation avoiding RPC.
-
- Note: If you want to create Connection with your own Configuration and NOT
use the RegionServer
- Connection (though its cache of locations will be warm, and its life-cycle is
not the concern
- of the CP), see createConnection(Configuration).
+ For light-weight usage only. Heavy-duty usage will pull down
+ the hosting RegionServer responsiveness as well as that of other Coprocessors
making use of
+ this Connection. Use to create table on start or to do administrative
operations. Coprocessors
+ should create their own Connections if heavy usage to avoid impinging on
hosting Server
+ operation. To create a Connection or if a Coprocessor requires a region with
a particular
+ Configuration, use ConnectionFactory or
+ createConnection(Configuration)}.
+
+ Be aware that operations that make use of this Connection are executed as
the RegionServer
+ User, the hbase super user that started this server process. Exercise caution
running
+ operations as this User (See createConnection(Configuration)}
to run as other than
+ the RegionServer User).
+
+ Be careful RPC'ing from a Coprocessor context. RPC's will fail, stall,
retry, and/or crawl
+ because the remote side is not online, is struggling or it is on the other
side of a network
+ partition. Any use of Connection from inside a Coprocessor must be able to
handle all such
+ hiccups.
Returns:
The host's Connection to the Cluster.
+See Also:
+createConnection(Configuration)
@@ -293,18 +304,26 @@ extends
createConnection
-ConnectioncreateConnection(org.apache.hadoop.conf.Configurationconf)
+ConnectioncreateConnection(org.apache.hadoop.conf.Configurationconf)
throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true;
title="class or interface in java.io">IOException
-Creates a cluster connection using the passed configuration.
- Using this Connection to get at a local resource -- say a Region that is
on the local
- Server or using Admin Interface from a Coprocessor hosted on the Master --
will result in a
- short-circuit of the RPC framework to make a direct invocation avoiding RPC.
-
- Note: HBase will NOT cache/maintain this Connection. If Coprocessors need to
cache and reuse
- this connection, it has to be done by Coprocessors.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
--
diff --git
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
index c0771ec..fa6927a 100644
---
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
+++
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
@@ -7,7 +7,7 @@
-
+
Apache HBase - Exemplar for hbase-client archetype
Dependency Information
@@ -147,7 +147,7 @@
https://www.apache.org/;>The Apache Software
Foundation.
All rights reserved.
- Last Published:
2017-12-05
+ Last Published:
2017-12-06
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
--
diff --git
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
index fdd16aa..3c1da71 100644
---
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
+++
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
@@ -7,7 +7,7 @@
-
+
Apache HBase - Exemplar for hbase-client archetype Project
Dependency Management
@@ -775,18 +775,24 @@
test-jar
https://www.apache.org/licenses/LICENSE-2.0.txt;>Apache License, Version
2.0
+org.apache.hbase
+http://hbase.apache.org/hbase-build-configuration/hbase-zookeeper;>hbase-zookeeper
+3.0.0-SNAPSHOT
+test-jar
+https://www.apache.org/licenses/LICENSE-2.0.txt;>Apache License, Version
2.0
+
org.bouncycastle
http://www.bouncycastle.org/java.html;>bcprov-jdk16
1.46
jar
http://www.bouncycastle.org/licence.html;>Bouncy Castle
Licence
-
+
org.hamcrest
https://github.com/hamcrest/JavaHamcrest/hamcrest-core;>hamcrest-core
1.3
jar
http://www.opensource.org/licenses/bsd-license.php;>New BSD
License
-
+
org.mockito
http://mockito.org;>mockito-core
2.1.0
@@ -804,7 +810,7 @@
https://www.apache.org/;>The Apache Software
Foundation.
All rights reserved.
- Last Published:
2017-12-05
+ Last Published:
2017-12-06
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html
--
diff --git
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html
index 2e2bdd1..943db8b 100644
--- a/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html
+++ b/hbase-build-configuration/hbase-archetypes/hbase-client-project/index.html
@@ -7,7 +7,7 @@
-
+
Apache HBase - Exemplar for hbase-client archetype
About
@@ -119,7 +119,7 @@
https://www.apache.org/;>The Apache Software
Foundation.
All rights reserved.
- Last Published:
2017-12-05
+ Last Published:
2017-12-06
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
--
diff --git
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
index 232443d..c65702d 100644
---
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
+++
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/integration.html
@@ -7,7 +7,7 @@
-
+
Apache HBase - Exemplar for hbase-client archetype CI
Management
@@ -126,7 +126,7 @@
https://www.apache.org/;>The Apache Software
Foundation.
All rights reserved.
- Last Published:
2017-12-05
+ Last Published:
2017-12-06
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/d171b896/hbase-build-configuration/hbase-archetypes/hbase-client-project/issue-tracking.html
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c54c242b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html
--
diff --git
a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html
b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html
index 698163a..238f6d6 100644
---
a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html
+++
b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html
@@ -7,7 +7,7 @@
-
+
Apache HBase - Exemplar for hbase-shaded-client archetype
Reactor Dependency Convergence
@@ -838,7 +838,7 @@
https://www.apache.org/;>The Apache Software
Foundation.
All rights reserved.
- Last Published:
2017-12-02
+ Last Published:
2017-12-03
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c54c242b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-info.html
--
diff --git
a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-info.html
b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-info.html
index f630509..d188ea3 100644
---
a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-info.html
+++
b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-info.html
@@ -7,7 +7,7 @@
-
+
Apache HBase - Exemplar for hbase-shaded-client archetype
Dependency Information
@@ -147,7 +147,7 @@
https://www.apache.org/;>The Apache Software
Foundation.
All rights reserved.
- Last Published:
2017-12-02
+ Last Published:
2017-12-03
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c54c242b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-management.html
--
diff --git
a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-management.html
b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-management.html
index faa52ff..41779ab 100644
---
a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-management.html
+++
b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-management.html
@@ -7,7 +7,7 @@
-
+
Apache HBase - Exemplar for hbase-shaded-client archetype
Project Dependency Management
@@ -804,7 +804,7 @@
https://www.apache.org/;>The Apache Software
Foundation.
All rights reserved.
- Last Published:
2017-12-02
+ Last Published:
2017-12-03
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c54c242b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/index.html
--
diff --git
a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/index.html
b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/index.html
index dc4de1d..e9a84e4 100644
---
a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/index.html
+++
b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/index.html
@@ -7,7 +7,7 @@
-
+
Apache HBase - Exemplar for hbase-shaded-client archetype
About
@@ -119,7 +119,7 @@
https://www.apache.org/;>The Apache Software
Foundation.
All rights reserved.
- Last Published:
2017-12-02
+ Last Published:
2017-12-03
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/c54c242b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/integration.html
--
diff --git
a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/integration.html
b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/integration.html
index c8d64af..2906ba6 100644
---
a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/integration.html
+++
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/713d773f/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.PBHelper.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.PBHelper.html
b/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.PBHelper.html
index 25e368d..d0f781f 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.PBHelper.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/io/asyncfs/FanOutOneBlockAsyncDFSOutputHelper.PBHelper.html
@@ -25,798 +25,798 @@
017 */
018package
org.apache.hadoop.hbase.io.asyncfs;
019
-020import static
org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelOption.CONNECT_TIMEOUT_MILLIS;
-021import static
org.apache.hadoop.hbase.shaded.io.netty.handler.timeout.IdleState.READER_IDLE;
-022import static
org.apache.hadoop.fs.CreateFlag.CREATE;
-023import static
org.apache.hadoop.fs.CreateFlag.OVERWRITE;
-024import static
org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createEncryptor;
-025import static
org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.trySaslNegotiate;
+020import static
org.apache.hadoop.fs.CreateFlag.CREATE;
+021import static
org.apache.hadoop.fs.CreateFlag.OVERWRITE;
+022import static
org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createEncryptor;
+023import static
org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.trySaslNegotiate;
+024import static
org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelOption.CONNECT_TIMEOUT_MILLIS;
+025import static
org.apache.hadoop.hbase.shaded.io.netty.handler.timeout.IdleState.READER_IDLE;
026import static
org.apache.hadoop.hdfs.DFSConfigKeys.DFS_CLIENT_SOCKET_TIMEOUT_KEY;
027import static
org.apache.hadoop.hdfs.DFSConfigKeys.DFS_CLIENT_USE_DN_HOSTNAME;
028import static
org.apache.hadoop.hdfs.DFSConfigKeys.DFS_CLIENT_USE_DN_HOSTNAME_DEFAULT;
029import static
org.apache.hadoop.hdfs.protocol.datatransfer.BlockConstructionStage.PIPELINE_SETUP_CREATE;
030
-031import
org.apache.hadoop.hbase.shaded.com.google.common.base.Throwables;
-032import
org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableMap;
-033import
com.google.protobuf.CodedOutputStream;
-034
-035import
org.apache.hadoop.hbase.shaded.io.netty.bootstrap.Bootstrap;
-036import
org.apache.hadoop.hbase.shaded.io.netty.buffer.ByteBuf;
-037import
org.apache.hadoop.hbase.shaded.io.netty.buffer.ByteBufAllocator;
-038import
org.apache.hadoop.hbase.shaded.io.netty.buffer.ByteBufOutputStream;
-039import
org.apache.hadoop.hbase.shaded.io.netty.buffer.PooledByteBufAllocator;
-040import
org.apache.hadoop.hbase.shaded.io.netty.channel.Channel;
-041import
org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelFuture;
-042import
org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelFutureListener;
-043import
org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandler;
-044import
org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelHandlerContext;
-045import
org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelInitializer;
-046import
org.apache.hadoop.hbase.shaded.io.netty.channel.ChannelPipeline;
-047import
org.apache.hadoop.hbase.shaded.io.netty.channel.EventLoop;
-048import
org.apache.hadoop.hbase.shaded.io.netty.channel.SimpleChannelInboundHandler;
-049import
org.apache.hadoop.hbase.shaded.io.netty.handler.codec.protobuf.ProtobufDecoder;
-050import
org.apache.hadoop.hbase.shaded.io.netty.handler.codec.protobuf.ProtobufVarint32FrameDecoder;
-051import
org.apache.hadoop.hbase.shaded.io.netty.handler.timeout.IdleStateEvent;
-052import
org.apache.hadoop.hbase.shaded.io.netty.handler.timeout.IdleStateHandler;
-053import
org.apache.hadoop.hbase.shaded.io.netty.util.concurrent.Future;
-054import
org.apache.hadoop.hbase.shaded.io.netty.util.concurrent.FutureListener;
-055import
org.apache.hadoop.hbase.shaded.io.netty.util.concurrent.Promise;
-056
-057import java.io.IOException;
-058import
java.lang.reflect.InvocationTargetException;
-059import java.lang.reflect.Method;
-060import java.util.ArrayList;
-061import java.util.EnumSet;
-062import java.util.List;
-063import java.util.concurrent.TimeUnit;
-064
-065import org.apache.commons.logging.Log;
-066import
org.apache.commons.logging.LogFactory;
-067import
org.apache.hadoop.conf.Configuration;
-068import
org.apache.hadoop.crypto.CryptoProtocolVersion;
-069import
org.apache.hadoop.crypto.Encryptor;
-070import org.apache.hadoop.fs.CreateFlag;
-071import org.apache.hadoop.fs.FileSystem;
-072import
org.apache.hadoop.fs.FileSystemLinkResolver;
-073import org.apache.hadoop.fs.Path;
-074import
org.apache.hadoop.fs.UnresolvedLinkException;
-075import
org.apache.hadoop.fs.permission.FsPermission;
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/fd365a2b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
index d438f22..7c59e27 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/client/HBaseAdmin.ModifyTableFuture.html
@@ -1290,8 +1290,8 @@
1282 CompactType
compactType) throws IOException {
1283switch (compactType) {
1284 case MOB:
-1285
compact(this.connection.getAdminForMaster(), getMobRegionInfo(tableName),
major,
-1286 columnFamily);
+1285
compact(this.connection.getAdminForMaster(),
RegionInfo.createMobRegionInfo(tableName),
+1286major, columnFamily);
1287break;
1288 case NORMAL:
1289checkTableExists(tableName);
@@ -3248,7 +3248,7 @@
3240 new
CallableAdminProtos.GetRegionInfoResponse.CompactionState() {
3241@Override
3242public
AdminProtos.GetRegionInfoResponse.CompactionState call() throws Exception {
-3243 RegionInfo info =
getMobRegionInfo(tableName);
+3243 RegionInfo info =
RegionInfo.createMobRegionInfo(tableName);
3244 GetRegionInfoRequest
request =
3245
RequestConverter.buildGetRegionInfoRequest(info.getRegionName(), true);
3246 GetRegionInfoResponse
response = masterAdmin.getRegionInfo(rpcController, request);
@@ -3312,7 +3312,7 @@
3304}
3305break;
3306 default:
-3307throw new
IllegalArgumentException("Unknowne compactType: " + compactType);
+3307throw new
IllegalArgumentException("Unknown compactType: " + compactType);
3308}
3309if (state != null) {
3310 return
ProtobufUtil.createCompactionState(state);
@@ -3847,325 +3847,320 @@
3839});
3840 }
3841
-3842 private RegionInfo
getMobRegionInfo(TableName tableName) {
-3843return
RegionInfoBuilder.newBuilder(tableName).setStartKey(Bytes.toBytes(".mob")).setRegionId(0)
-3844.build();
-3845 }
-3846
-3847 private RpcControllerFactory
getRpcControllerFactory() {
-3848return this.rpcControllerFactory;
-3849 }
-3850
-3851 @Override
-3852 public void addReplicationPeer(String
peerId, ReplicationPeerConfig peerConfig, boolean enabled)
-3853 throws IOException {
-3854executeCallable(new
MasterCallableVoid(getConnection(), getRpcControllerFactory()) {
-3855 @Override
-3856 protected Void rpcCall() throws
Exception {
-3857
master.addReplicationPeer(getRpcController(),
-3858
RequestConverter.buildAddReplicationPeerRequest(peerId, peerConfig,
enabled));
-3859return null;
-3860 }
-3861});
-3862 }
-3863
-3864 @Override
-3865 public void
removeReplicationPeer(String peerId) throws IOException {
-3866executeCallable(new
MasterCallableVoid(getConnection(), getRpcControllerFactory()) {
-3867 @Override
-3868 protected Void rpcCall() throws
Exception {
-3869
master.removeReplicationPeer(getRpcController(),
-3870
RequestConverter.buildRemoveReplicationPeerRequest(peerId));
-3871return null;
-3872 }
-3873});
-3874 }
-3875
-3876 @Override
-3877 public void
enableReplicationPeer(final String peerId) throws IOException {
-3878executeCallable(new
MasterCallableVoid(getConnection(), getRpcControllerFactory()) {
-3879 @Override
-3880 protected Void rpcCall() throws
Exception {
-3881
master.enableReplicationPeer(getRpcController(),
-3882
RequestConverter.buildEnableReplicationPeerRequest(peerId));
-3883return null;
-3884 }
-3885});
-3886 }
-3887
-3888 @Override
-3889 public void
disableReplicationPeer(final String peerId) throws IOException {
-3890executeCallable(new
MasterCallableVoid(getConnection(), getRpcControllerFactory()) {
-3891 @Override
-3892 protected Void rpcCall() throws
Exception {
-3893
master.disableReplicationPeer(getRpcController(),
-3894
RequestConverter.buildDisableReplicationPeerRequest(peerId));
-3895return null;
-3896 }
-3897});
-3898 }
-3899
-3900 @Override
-3901 public ReplicationPeerConfig
getReplicationPeerConfig(final String peerId) throws IOException {
-3902return executeCallable(new
MasterCallableReplicationPeerConfig(getConnection(),
-3903getRpcControllerFactory()) {
-3904 @Override
-3905 protected ReplicationPeerConfig
rpcCall() throws Exception {
-3906GetReplicationPeerConfigResponse
response = master.getReplicationPeerConfig(
-3907 getRpcController(),
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b9722a17/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
--
diff --git
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
index 29ea7b3..6ed75c9 100644
---
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
+++
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ReplayBatchOperation.html
@@ -1313,7093 +1313,7082 @@
1305
1306 @Override
1307 public boolean isSplittable() {
-1308boolean result = isAvailable()
!hasReferences();
-1309LOG.info("ASKED IF SPLITTABLE " +
result + " " + getRegionInfo().getShortNameToLog(),
-1310 new Throwable("LOGGING:
REMOVE"));
-1311// REMOVE BELOW
-1312LOG.info("DEBUG LIST ALL FILES");
-1313for (HStore store :
this.stores.values()) {
-1314 LOG.info("store " +
store.getColumnFamilyName());
-1315 for (HStoreFile sf :
store.getStorefiles()) {
-1316
LOG.info(sf.toStringDetailed());
-1317 }
-1318}
-1319return result;
-1320 }
-1321
-1322 @Override
-1323 public boolean isMergeable() {
-1324if (!isAvailable()) {
-1325 LOG.debug("Region " + this
-1326 + " is not mergeable because
it is closing or closed");
-1327 return false;
-1328}
-1329if (hasReferences()) {
-1330 LOG.debug("Region " + this
-1331 + " is not mergeable because
it has references");
-1332 return false;
-1333}
-1334
-1335return true;
+1308return isAvailable()
!hasReferences();
+1309 }
+1310
+1311 @Override
+1312 public boolean isMergeable() {
+1313if (!isAvailable()) {
+1314 LOG.debug("Region " + this
+1315 + " is not mergeable because
it is closing or closed");
+1316 return false;
+1317}
+1318if (hasReferences()) {
+1319 LOG.debug("Region " + this
+1320 + " is not mergeable because
it has references");
+1321 return false;
+1322}
+1323
+1324return true;
+1325 }
+1326
+1327 public boolean areWritesEnabled() {
+1328synchronized(this.writestate) {
+1329 return
this.writestate.writesEnabled;
+1330}
+1331 }
+1332
+1333 @VisibleForTesting
+1334 public MultiVersionConcurrencyControl
getMVCC() {
+1335return mvcc;
1336 }
1337
-1338 public boolean areWritesEnabled() {
-1339synchronized(this.writestate) {
-1340 return
this.writestate.writesEnabled;
-1341}
-1342 }
-1343
-1344 @VisibleForTesting
-1345 public MultiVersionConcurrencyControl
getMVCC() {
-1346return mvcc;
-1347 }
-1348
-1349 @Override
-1350 public long getMaxFlushedSeqId() {
-1351return maxFlushedSeqId;
+1338 @Override
+1339 public long getMaxFlushedSeqId() {
+1340return maxFlushedSeqId;
+1341 }
+1342
+1343 /**
+1344 * @return readpoint considering given
IsolationLevel. Pass {@code null} for default
+1345 */
+1346 public long
getReadPoint(IsolationLevel isolationLevel) {
+1347if (isolationLevel != null
isolationLevel == IsolationLevel.READ_UNCOMMITTED) {
+1348 // This scan can read even
uncommitted transactions
+1349 return Long.MAX_VALUE;
+1350}
+1351return mvcc.getReadPoint();
1352 }
1353
-1354 /**
-1355 * @return readpoint considering given
IsolationLevel. Pass {@code null} for default
-1356 */
-1357 public long
getReadPoint(IsolationLevel isolationLevel) {
-1358if (isolationLevel != null
isolationLevel == IsolationLevel.READ_UNCOMMITTED) {
-1359 // This scan can read even
uncommitted transactions
-1360 return Long.MAX_VALUE;
-1361}
-1362return mvcc.getReadPoint();
-1363 }
-1364
-1365 public boolean
isLoadingCfsOnDemandDefault() {
-1366return
this.isLoadingCfsOnDemandDefault;
-1367 }
-1368
-1369 /**
-1370 * Close down this HRegion. Flush the
cache, shut down each HStore, don't
-1371 * service any more calls.
-1372 *
-1373 * pThis method could take
some time to execute, so don't call it from a
-1374 * time-sensitive thread.
-1375 *
-1376 * @return Vector of all the storage
files that the HRegion's component
-1377 * HStores make use of. It's a list
of all StoreFile objects. Returns empty
-1378 * vector if already closed and null
if judged that it should not close.
-1379 *
-1380 * @throws IOException e
-1381 * @throws DroppedSnapshotException
Thrown when replay of wal is required
-1382 * because a Snapshot was not properly
persisted. The region is put in closing mode, and the
-1383 * caller MUST abort after this.
-1384 */
-1385 public Mapbyte[],
ListHStoreFile close() throws IOException {
-1386return close(false);
-1387 }
-1388
-1389 private final Object closeLock = new
Object();
-1390
-1391 /** Conf key for the periodic flush
1 - 100 of 298 matches
Mail list logo