hbase git commit: HBASE-18997 Remove the redundant methods in RegionInfo

2017-10-14 Thread chia7712
Repository: hbase
Updated Branches:
  refs/heads/master 5d9b7a978 -> 240b4b16f


HBASE-18997 Remove the redundant methods in RegionInfo


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/240b4b16
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/240b4b16
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/240b4b16

Branch: refs/heads/master
Commit: 240b4b16ff14122bb84101d81d5bb263ea7925f3
Parents: 5d9b7a9
Author: Chia-Ping Tsai 
Authored: Sat Oct 14 23:26:11 2017 +0800
Committer: Chia-Ping Tsai 
Committed: Sat Oct 14 23:26:11 2017 +0800

--
 .../hadoop/hbase/client/AsyncRegionLocator.java |  2 +-
 .../apache/hadoop/hbase/client/RegionInfo.java  | 10 
 .../hadoop/hbase/client/RegionInfoBuilder.java  | 16 -
 .../TestImportTSVWithOperationAttributes.java   |  2 +-
 .../hbase/mapreduce/TestImportTSVWithTTLs.java  |  2 +-
 .../hbase/favored/FavoredNodesManager.java  |  2 +-
 .../org/apache/hadoop/hbase/master/HMaster.java |  2 +-
 .../master/assignment/AssignmentManager.java|  6 ++---
 .../hbase/master/balancer/BaseLoadBalancer.java |  2 +-
 .../master/balancer/RegionLocationFinder.java   |  2 +-
 .../hbase/mob/DefaultMobStoreFlusher.java   |  2 +-
 .../AnnotationReadingPriorityFunction.java  |  4 ++--
 .../hadoop/hbase/regionserver/HRegion.java  |  5 ++--
 .../hbase/regionserver/HRegionServer.java   |  6 ++---
 .../hbase/regionserver/RSRpcServices.java   | 10 
 .../hadoop/hbase/regionserver/StoreFlusher.java |  2 +-
 .../regionserver/StorefileRefresherChore.java   |  2 +-
 .../org/apache/hadoop/hbase/util/HBaseFsck.java |  2 +-
 .../apache/hadoop/hbase/client/TestAdmin2.java  |  4 ++--
 ...syncNonMetaRegionLocatorConcurrenyLimit.java |  4 ++--
 .../hbase/client/TestAsyncRegionAdminApi.java   |  4 ++--
 .../hbase/coprocessor/TestMasterObserver.java   |  2 +-
 .../master/TestDistributedLogSplitting.java |  6 ++---
 .../hbase/regionserver/TestHRegionInfo.java |  2 +-
 .../hbase/regionserver/TestPriorityRpc.java | 24 
 .../regionserver/TestRegionInfoBuilder.java |  2 +-
 .../hadoop/hbase/util/BaseTestHBaseFsck.java|  2 +-
 27 files changed, 53 insertions(+), 76 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/240b4b16/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncRegionLocator.java
--
diff --git 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncRegionLocator.java
 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncRegionLocator.java
index 2be75dc..6a28708 100644
--- 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncRegionLocator.java
+++ 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncRegionLocator.java
@@ -142,7 +142,7 @@ class AsyncRegionLocator {
   }
 
   void updateCachedLocation(HRegionLocation loc, Throwable exception) {
-if (loc.getRegionInfo().isMetaTable()) {
+if (loc.getRegion().isMetaRegion()) {
   metaRegionLocator.updateCachedLocation(loc, exception);
 } else {
   nonMetaRegionLocator.updateCachedLocation(loc, exception);

http://git-wip-us.apache.org/repos/asf/hbase/blob/240b4b16/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfo.java
--
diff --git 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfo.java 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfo.java
index e9716e9..0eb4e42 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfo.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfo.java
@@ -215,16 +215,6 @@ public interface RegionInfo {
   boolean isSplitParent();
 
   /**
-   * @return true if this region is from hbase:meta.
-   */
-  boolean isMetaTable();
-
-  /**
-   * @return true if this region is from a system table.
-   */
-  boolean isSystemTable();
-
-  /**
* @return true if this region is a meta region.
*/
   boolean isMetaRegion();

http://git-wip-us.apache.org/repos/asf/hbase/blob/240b4b16/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfoBuilder.java
--
diff --git 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfoBuilder.java
 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfoBuilder.java
index acff186..68f839a 100644
--- 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfoBuilder.java
+++ 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfoBuilder.java
@@ -362,14 +362,6 @@ public class RegionInfoBuilder {
  Bytes.equals(endKey, HCo

hbase git commit: HBASE-18997 Remove the redundant methods in RegionInfo

2017-10-14 Thread chia7712
Repository: hbase
Updated Branches:
  refs/heads/branch-2 a63d79837 -> d2e4e0e1f


HBASE-18997 Remove the redundant methods in RegionInfo


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/d2e4e0e1
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/d2e4e0e1
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/d2e4e0e1

Branch: refs/heads/branch-2
Commit: d2e4e0e1f397505b051712393c4012b056fc718e
Parents: a63d798
Author: Chia-Ping Tsai 
Authored: Sat Oct 14 23:26:11 2017 +0800
Committer: Chia-Ping Tsai 
Committed: Sat Oct 14 23:32:48 2017 +0800

--
 .../hadoop/hbase/client/AsyncRegionLocator.java |  2 +-
 .../apache/hadoop/hbase/client/RegionInfo.java  | 10 
 .../hadoop/hbase/client/RegionInfoBuilder.java  | 16 -
 .../TestImportTSVWithOperationAttributes.java   |  2 +-
 .../hbase/mapreduce/TestImportTSVWithTTLs.java  |  2 +-
 .../hbase/favored/FavoredNodesManager.java  |  2 +-
 .../org/apache/hadoop/hbase/master/HMaster.java |  2 +-
 .../master/assignment/AssignmentManager.java|  6 ++---
 .../hbase/master/balancer/BaseLoadBalancer.java |  2 +-
 .../master/balancer/RegionLocationFinder.java   |  2 +-
 .../hbase/mob/DefaultMobStoreFlusher.java   |  2 +-
 .../AnnotationReadingPriorityFunction.java  |  4 ++--
 .../hadoop/hbase/regionserver/HRegion.java  |  5 ++--
 .../hbase/regionserver/HRegionServer.java   |  6 ++---
 .../hbase/regionserver/RSRpcServices.java   | 10 
 .../hadoop/hbase/regionserver/StoreFlusher.java |  2 +-
 .../regionserver/StorefileRefresherChore.java   |  2 +-
 .../org/apache/hadoop/hbase/util/HBaseFsck.java |  2 +-
 .../apache/hadoop/hbase/client/TestAdmin2.java  |  4 ++--
 ...syncNonMetaRegionLocatorConcurrenyLimit.java |  4 ++--
 .../hbase/client/TestAsyncRegionAdminApi.java   |  4 ++--
 .../hbase/coprocessor/TestMasterObserver.java   |  2 +-
 .../master/TestDistributedLogSplitting.java |  6 ++---
 .../hbase/regionserver/TestHRegionInfo.java |  2 +-
 .../hbase/regionserver/TestPriorityRpc.java | 24 
 .../regionserver/TestRegionInfoBuilder.java |  2 +-
 .../hadoop/hbase/util/BaseTestHBaseFsck.java|  2 +-
 27 files changed, 53 insertions(+), 76 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/d2e4e0e1/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncRegionLocator.java
--
diff --git 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncRegionLocator.java
 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncRegionLocator.java
index 2be75dc..6a28708 100644
--- 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncRegionLocator.java
+++ 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncRegionLocator.java
@@ -142,7 +142,7 @@ class AsyncRegionLocator {
   }
 
   void updateCachedLocation(HRegionLocation loc, Throwable exception) {
-if (loc.getRegionInfo().isMetaTable()) {
+if (loc.getRegion().isMetaRegion()) {
   metaRegionLocator.updateCachedLocation(loc, exception);
 } else {
   nonMetaRegionLocator.updateCachedLocation(loc, exception);

http://git-wip-us.apache.org/repos/asf/hbase/blob/d2e4e0e1/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfo.java
--
diff --git 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfo.java 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfo.java
index e9716e9..0eb4e42 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfo.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfo.java
@@ -215,16 +215,6 @@ public interface RegionInfo {
   boolean isSplitParent();
 
   /**
-   * @return true if this region is from hbase:meta.
-   */
-  boolean isMetaTable();
-
-  /**
-   * @return true if this region is from a system table.
-   */
-  boolean isSystemTable();
-
-  /**
* @return true if this region is a meta region.
*/
   boolean isMetaRegion();

http://git-wip-us.apache.org/repos/asf/hbase/blob/d2e4e0e1/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfoBuilder.java
--
diff --git 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfoBuilder.java
 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfoBuilder.java
index acff186..68f839a 100644
--- 
a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfoBuilder.java
+++ 
b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RegionInfoBuilder.java
@@ -362,14 +362,6 @@ public class RegionInfoBuilder {
  Bytes.equals(endKey,

hbase git commit: HBASE-18966 Use non-sync TimeRangeTracker as a replacement for TimeRange in ImmutableSegment

2017-10-14 Thread chia7712
Repository: hbase
Updated Branches:
  refs/heads/master d653c0bc7 -> 5d9b7a978


HBASE-18966 Use non-sync TimeRangeTracker as a replacement for TimeRange in 
ImmutableSegment


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/5d9b7a97
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/5d9b7a97
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/5d9b7a97

Branch: refs/heads/master
Commit: 5d9b7a978cf84ed8aed91e93fcc6ddbed50cf47c
Parents: d653c0b
Author: Chia-Ping Tsai 
Authored: Sat Oct 14 23:09:58 2017 +0800
Committer: Chia-Ping Tsai 
Committed: Sat Oct 14 23:09:58 2017 +0800

--
 .../org/apache/hadoop/hbase/util/ClassSize.java |  5 ++
 .../regionserver/CompositeImmutableSegment.java | 11 ---
 .../hbase/regionserver/ImmutableSegment.java| 27 +--
 .../hbase/regionserver/MutableSegment.java  | 17 +---
 .../hadoop/hbase/regionserver/Segment.java  | 21 ++---
 .../hbase/regionserver/TimeRangeTracker.java|  4 +-
 .../apache/hadoop/hbase/io/TestHeapSize.java| 32 
 .../TestCompactingToCellFlatMapMemStore.java| 83 
 .../hadoop/hbase/regionserver/TestHStore.java   | 10 ++-
 9 files changed, 133 insertions(+), 77 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/5d9b7a97/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
index a4d5792..d9ea761 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
@@ -132,6 +132,9 @@ public class ClassSize {
   /** Overhead for SyncTimeRangeTracker */
   public static final int SYNC_TIMERANGE_TRACKER;
 
+  /** Overhead for NonSyncTimeRangeTracker */
+  public static final int NON_SYNC_TIMERANGE_TRACKER;
+
   /** Overhead for CellSkipListSet */
   public static final int CELL_SET;
 
@@ -327,6 +330,8 @@ public class ClassSize {
 TIMERANGE = align(ClassSize.OBJECT + Bytes.SIZEOF_LONG * 2 + 
Bytes.SIZEOF_BOOLEAN);
 
 SYNC_TIMERANGE_TRACKER = align(ClassSize.OBJECT + 2 * REFERENCE);
+NON_SYNC_TIMERANGE_TRACKER = align(ClassSize.OBJECT + 2 * 
Bytes.SIZEOF_LONG);
+
 CELL_SET = align(OBJECT + REFERENCE);
 
 STORE_SERVICES = align(OBJECT + REFERENCE + ATOMIC_LONG);

http://git-wip-us.apache.org/repos/asf/hbase/blob/5d9b7a97/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompositeImmutableSegment.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompositeImmutableSegment.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompositeImmutableSegment.java
index 0d2608f..ef4c3cd 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompositeImmutableSegment.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompositeImmutableSegment.java
@@ -42,17 +42,11 @@ import 
org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTe
 public class CompositeImmutableSegment extends ImmutableSegment {
 
   private final List segments;
-  // CompositeImmutableSegment is used for snapshots and snapshot should
-  // support getTimeRangeTracker() interface.
-  // Thus we hold a constant TRT build in the construction time from TRT of 
the given segments.
-  private final TimeRangeTracker timeRangeTracker;
-
   private long keySize = 0;
 
   public CompositeImmutableSegment(CellComparator comparator, 
List segments) {
 super(comparator);
 this.segments = segments;
-this.timeRangeTracker = 
TimeRangeTracker.create(TimeRangeTracker.Type.SYNC);
 for (ImmutableSegment s : segments) {
   this.timeRangeTracker.includeTimestamp(s.getTimeRangeTracker().getMax());
   this.timeRangeTracker.includeTimestamp(s.getTimeRangeTracker().getMin());
@@ -127,11 +121,6 @@ public class CompositeImmutableSegment extends 
ImmutableSegment {
 throw new IllegalStateException("Not supported by 
CompositeImmutableScanner");
   }
 
-  @Override
-  public long getMinTimestamp(){
-throw new IllegalStateException("Not supported by 
CompositeImmutableScanner");
-  }
-
   /**
* Creates the scanner for the given read point
* @return a scanner for the given read point

http://git-wip-us.apache.org/repos/asf/hbase/blob/5d9b7a97/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ImmutableSegment.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/

hbase git commit: HBASE-18966 Use non-sync TimeRangeTracker as a replacement for TimeRange in ImmutableSegment

2017-10-14 Thread chia7712
Repository: hbase
Updated Branches:
  refs/heads/branch-2 104595137 -> a63d79837


HBASE-18966 Use non-sync TimeRangeTracker as a replacement for TimeRange in 
ImmutableSegment


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/a63d7983
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/a63d7983
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/a63d7983

Branch: refs/heads/branch-2
Commit: a63d79837d2c47c2b226bd5d777f7beaedf95a8d
Parents: 1045951
Author: Chia-Ping Tsai 
Authored: Sat Oct 14 23:09:58 2017 +0800
Committer: Chia-Ping Tsai 
Committed: Sat Oct 14 23:16:48 2017 +0800

--
 .../org/apache/hadoop/hbase/util/ClassSize.java |  5 ++
 .../regionserver/CompositeImmutableSegment.java | 11 ---
 .../hbase/regionserver/ImmutableSegment.java| 27 +--
 .../hbase/regionserver/MutableSegment.java  | 17 +---
 .../hadoop/hbase/regionserver/Segment.java  | 21 ++---
 .../hbase/regionserver/TimeRangeTracker.java|  4 +-
 .../apache/hadoop/hbase/io/TestHeapSize.java| 32 
 .../TestCompactingToCellFlatMapMemStore.java| 83 
 .../hadoop/hbase/regionserver/TestHStore.java   | 10 ++-
 9 files changed, 133 insertions(+), 77 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/a63d7983/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
index a4d5792..d9ea761 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java
@@ -132,6 +132,9 @@ public class ClassSize {
   /** Overhead for SyncTimeRangeTracker */
   public static final int SYNC_TIMERANGE_TRACKER;
 
+  /** Overhead for NonSyncTimeRangeTracker */
+  public static final int NON_SYNC_TIMERANGE_TRACKER;
+
   /** Overhead for CellSkipListSet */
   public static final int CELL_SET;
 
@@ -327,6 +330,8 @@ public class ClassSize {
 TIMERANGE = align(ClassSize.OBJECT + Bytes.SIZEOF_LONG * 2 + 
Bytes.SIZEOF_BOOLEAN);
 
 SYNC_TIMERANGE_TRACKER = align(ClassSize.OBJECT + 2 * REFERENCE);
+NON_SYNC_TIMERANGE_TRACKER = align(ClassSize.OBJECT + 2 * 
Bytes.SIZEOF_LONG);
+
 CELL_SET = align(OBJECT + REFERENCE);
 
 STORE_SERVICES = align(OBJECT + REFERENCE + ATOMIC_LONG);

http://git-wip-us.apache.org/repos/asf/hbase/blob/a63d7983/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompositeImmutableSegment.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompositeImmutableSegment.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompositeImmutableSegment.java
index 0d2608f..ef4c3cd 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompositeImmutableSegment.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompositeImmutableSegment.java
@@ -42,17 +42,11 @@ import 
org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTe
 public class CompositeImmutableSegment extends ImmutableSegment {
 
   private final List segments;
-  // CompositeImmutableSegment is used for snapshots and snapshot should
-  // support getTimeRangeTracker() interface.
-  // Thus we hold a constant TRT build in the construction time from TRT of 
the given segments.
-  private final TimeRangeTracker timeRangeTracker;
-
   private long keySize = 0;
 
   public CompositeImmutableSegment(CellComparator comparator, 
List segments) {
 super(comparator);
 this.segments = segments;
-this.timeRangeTracker = 
TimeRangeTracker.create(TimeRangeTracker.Type.SYNC);
 for (ImmutableSegment s : segments) {
   this.timeRangeTracker.includeTimestamp(s.getTimeRangeTracker().getMax());
   this.timeRangeTracker.includeTimestamp(s.getTimeRangeTracker().getMin());
@@ -127,11 +121,6 @@ public class CompositeImmutableSegment extends 
ImmutableSegment {
 throw new IllegalStateException("Not supported by 
CompositeImmutableScanner");
   }
 
-  @Override
-  public long getMinTimestamp(){
-throw new IllegalStateException("Not supported by 
CompositeImmutableScanner");
-  }
-
   /**
* Creates the scanner for the given read point
* @return a scanner for the given read point

http://git-wip-us.apache.org/repos/asf/hbase/blob/a63d7983/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ImmutableSegment.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionser

[26/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/KeyValueHeap.KVScannerComparator.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/KeyValueHeap.KVScannerComparator.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/KeyValueHeap.KVScannerComparator.html
index 35716ee..258f095 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/KeyValueHeap.KVScannerComparator.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/KeyValueHeap.KVScannerComparator.html
@@ -148,317 +148,311 @@
 140   * 

141 * This can ONLY be called when you are using Scanners that implement InternalScanner as well as 142 * KeyValueScanner (a {@link StoreScanner}). -143 * @param result -144 * @return true if more rows exist after this one, false if scanner is done -145 */ -146 @Override -147 public boolean next(List result) throws IOException { -148return next(result, NoLimitScannerContext.getInstance()); -149 } -150 -151 @Override -152 public boolean next(List result, ScannerContext scannerContext) throws IOException { -153if (this.current == null) { -154 return scannerContext.setScannerState(NextState.NO_MORE_VALUES).hasMoreValues(); -155} -156InternalScanner currentAsInternal = (InternalScanner)this.current; -157boolean moreCells = currentAsInternal.next(result, scannerContext); -158Cell pee = this.current.peek(); -159 -160/* -161 * By definition, any InternalScanner must return false only when it has no -162 * further rows to be fetched. So, we can close a scanner if it returns -163 * false. All existing implementations seem to be fine with this. It is much -164 * more efficient to close scanners which are not needed than keep them in -165 * the heap. This is also required for certain optimizations. -166 */ -167 -168if (pee == null || !moreCells) { -169 // add the scanner that is to be closed -170 this.scannersForDelayedClose.add(this.current); -171} else { -172 this.heap.add(this.current); -173} -174this.current = null; -175this.current = pollRealKV(); -176if (this.current == null) { -177 moreCells = scannerContext.setScannerState(NextState.NO_MORE_VALUES).hasMoreValues(); -178} -179return moreCells; -180 } -181 -182 protected static class KVScannerComparator implements Comparator { -183protected CellComparator kvComparator; -184/** -185 * Constructor -186 * @param kvComparator -187 */ -188public KVScannerComparator(CellComparator kvComparator) { -189 this.kvComparator = kvComparator; -190} -191public int compare(KeyValueScanner left, KeyValueScanner right) { -192 int comparison = compare(left.peek(), right.peek()); -193 if (comparison != 0) { -194return comparison; -195 } else { -196// Since both the keys are exactly the same, we break the tie in favor of higher ordered -197// scanner since it'll have newer data. Since higher value should come first, we reverse -198// sort here. -199return Long.compare(right.getScannerOrder(), left.getScannerOrder()); -200 } -201} -202/** -203 * Compares two KeyValue -204 * @param left -205 * @param right -206 * @return less than 0 if left is smaller, 0 if equal etc.. +143 * @return true if more rows exist after this one, false if scanner is done +144 */ +145 @Override +146 public boolean next(List result, ScannerContext scannerContext) throws IOException { +147if (this.current == null) { +148 return scannerContext.setScannerState(NextState.NO_MORE_VALUES).hasMoreValues(); +149} +150InternalScanner currentAsInternal = (InternalScanner)this.current; +151boolean moreCells = currentAsInternal.next(result, scannerContext); +152Cell pee = this.current.peek(); +153 +154/* +155 * By definition, any InternalScanner must return false only when it has no +156 * further rows to be fetched. So, we can close a scanner if it returns +157 * false. All existing implementations seem to be fine with this. It is much +158 * more efficient to close scanners which are not needed than keep them in +159 * the heap. This is also required for certain optimizations. +160 */ +161 +162if (pee == null || !moreCells) { +163 // add the scanner that is to be closed +164 this.scannersForDelayedClose.add(this.current); +165} else { +166 this.heap.add(this.current); +167} +168this.current = null; +169this.current = pollRealKV(); +170if (this.current == null) { +171 moreCells = scannerContext.setScannerState(NextState.NO_MORE_VALUES).hasMoreValues(); +172} +173return moreCells; +174 } +175 +176


[23/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/export_control.html
--
diff --git a/export_control.html b/export_control.html
index 46d070e..dd3c6c0 100644
--- a/export_control.html
+++ b/export_control.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – 
   Export Control
@@ -336,7 +336,7 @@ for more details.
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-10-13
+  Last Published: 
2017-10-14
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-annotations/dependencies.html
--
diff --git a/hbase-annotations/dependencies.html 
b/hbase-annotations/dependencies.html
index c34ffab..2b19b94 100644
--- a/hbase-annotations/dependencies.html
+++ b/hbase-annotations/dependencies.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Annotations – Project Dependencies
 
@@ -272,7 +272,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-10-13
+  Last Published: 
2017-10-14
 
 
 



[04/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/testdevapidocs/org/apache/hadoop/hbase/snapshot/TestExportSnapshot.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/snapshot/TestExportSnapshot.html 
b/testdevapidocs/org/apache/hadoop/hbase/snapshot/TestExportSnapshot.html
index a5e263a..86ef89a 100644
--- a/testdevapidocs/org/apache/hadoop/hbase/snapshot/TestExportSnapshot.html
+++ b/testdevapidocs/org/apache/hadoop/hbase/snapshot/TestExportSnapshot.html
@@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab";
 
 
 
-public class TestExportSnapshot
+public class TestExportSnapshot
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
 Test Export Snapshot Tool
 
@@ -389,7 +389,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 timeout
-public final org.junit.rules.TestRule timeout
+public final org.junit.rules.TestRule timeout
 
 
 
@@ -398,7 +398,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 LOG
-private static final org.apache.commons.logging.Log LOG
+private static final org.apache.commons.logging.Log LOG
 
 
 
@@ -407,7 +407,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 TEST_UTIL
-protected static final HBaseTestingUtility TEST_UTIL
+protected static final HBaseTestingUtility TEST_UTIL
 
 
 
@@ -416,7 +416,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 FAMILY
-protected static final byte[] FAMILY
+protected static final byte[] FAMILY
 
 
 
@@ -425,7 +425,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 testName
-public final org.junit.rules.TestName testName
+public final org.junit.rules.TestName testName
 
 
 
@@ -434,7 +434,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 tableName
-protected org.apache.hadoop.hbase.TableName tableName
+protected org.apache.hadoop.hbase.TableName tableName
 
 
 
@@ -443,7 +443,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 emptySnapshotName
-private byte[] emptySnapshotName
+private byte[] emptySnapshotName
 
 
 
@@ -452,7 +452,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 snapshotName
-private byte[] snapshotName
+private byte[] snapshotName
 
 
 
@@ -461,7 +461,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 tableNumFiles
-private int tableNumFiles
+private int tableNumFiles
 
 
 
@@ -470,7 +470,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 admin
-private org.apache.hadoop.hbase.client.Admin admin
+private org.apache.hadoop.hbase.client.Admin admin
 
 
 
@@ -487,7 +487,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 TestExportSnapshot
-public TestExportSnapshot()
+public TestExportSnapshot()
 
 
 
@@ -504,7 +504,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 setUpBaseConf
-public static void setUpBaseConf(org.apache.hadoop.conf.Configuration conf)
+public static void setUpBaseConf(org.apache.hadoop.conf.Configuration conf)
 
 
 
@@ -513,7 +513,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 setUpBeforeClass
-public static void setUpBeforeClass()
+public static void setUpBeforeClass()
  throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true";
 title="class or interface in java.lang">Exception
 
 Throws:
@@ -527,7 +527,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 tearDownAfterClass
-public static void tearDownAfterClass()
+public static void tearDownAfterClass()
throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true";
 title="class or interface in java.lang">Exception
 
 Throws:
@@ -541,7 +541,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 setUp
-public void setUp()
+public void setUp()
throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true";
 title="class or interface in java.lang">Exception
 Create a table and take a snapshot of the table used by the 
export test.
 
@@ -556,7 +556,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 createTable
-protected void createTable()
+protected void createTable()
 throws http://docs.oracle.com/javase/8/docs/api/java/lang/Exception.html?is-external=true";
 title="class or interface in java.lang">Exception
 
 Throws:
@@ -570,7 +570,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 getBypassRegionPredicate
-protected TestExportSnapshot.RegionPredicate getBypassRegionPredicate()
+protect

[18/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-build-configuration/hbase-archetypes/dependency-convergence.html
--
diff --git 
a/hbase-build-configuration/hbase-archetypes/dependency-convergence.html 
b/hbase-build-configuration/hbase-archetypes/dependency-convergence.html
index adc4525..cdf7fa0 100644
--- a/hbase-build-configuration/hbase-archetypes/dependency-convergence.html
+++ b/hbase-build-configuration/hbase-archetypes/dependency-convergence.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Archetypes – Reactor Dependency 
Convergence
 
@@ -123,10 +123,10 @@
 305
 
 Number of unique artifacts (NOA):
-331
+329
 
 Number of version-conflicting artifacts (NOC):
-17
+16
 
 Number of SNAPSHOT artifacts (NOS):
 0
@@ -191,75 +191,20 @@
 11.0.2
 
 
-org.apache.hbase:hbase-assembly:pom:3.0.0-SNAPSHOT+- org.apache.hbase:hbase-server:jar:3.0.0-SNAPSHOT:compile|  \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)\- org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-common:jar:2.7.1:compile|  \- (com.google.guava:guava:jar:11.0.2:compile - omitted for 
duplicate)\- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-client-project:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-client:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-endpoint:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-examples:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-mapreduce-client-core:jar:2.7.1:compile|  \- org.apache.hadoop:hadoop-yarn-common:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-hadoop-compat:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-metrics-api:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile  \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-hadoop2-compat:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-it:jar:3.0.0-SNAPSHOT+- org.apache.hbase:hbase-mapreduce:jar:3.0.0-SNAPSHOT:compile|  \- org.apache.hadoop:hadoop-hdfs:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)+- org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT:compile|  \- com.google.guava:guava:jar:11.0.2:compile\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- (com.google.guava:guava:jar:11.0.2:compile - 
omitted for duplicate)
-org.apache.hbase:hbase-mapreduce:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-common:jar:2.7.1:compile|  \- com.google.guava:guava:jar:11.0.2:compile+- org.apache.hadoop:hadoop-hdfs:jar:2.7.1:compile|  \- (com.google.guava:guava:jar:11.0.2:compile - omitted for 
duplicate)\- org.apache.hadoop:hadoop-minicluster:jar:2.7.1:test   +- org.apache.hadoop:hadoop-common:test-jar:tests:2.7.1:test   |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)   \- org.apache.hadoop:hadoop-yarn-server-tests:test-jar:tests:2.7.1:test  +- org.apache.hadoop:hadoop-yarn-server-nodemanager:jar:2.7.1:test  |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  +- org
 .apache.hadoop:hadoop-yarn-server-resourcemanager:jar:2.7.1:test  |  +- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  |  +- org.apache.hadoop:hadoop-yarn-server-applicationhistoryservice:jar:2.7.1:test  |  |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  |  \- org.apache.hadoop:hadoop-yarn-server-web-proxy:jar:2.7.1:test  | \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)
-org.apache.hbase:hbase-metrics-api:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-metrics:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-commo

hbase-site git commit: INFRA-10751 Empty commit

2017-10-14 Thread git-site-role
Repository: hbase-site
Updated Branches:
  refs/heads/asf-site 7816cbdea -> 6ca7e9400


INFRA-10751 Empty commit


Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/6ca7e940
Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/6ca7e940
Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/6ca7e940

Branch: refs/heads/asf-site
Commit: 6ca7e9400f5ddedeb31e4bbc6af21f73bbcc9597
Parents: 7816cbd
Author: jenkins 
Authored: Sat Oct 14 15:16:03 2017 +
Committer: jenkins 
Committed: Sat Oct 14 15:16:03 2017 +

--

--




[33/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupCopyJob.BackupDistCp.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupCopyJob.BackupDistCp.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupCopyJob.BackupDistCp.html
index 65a31ce..7551292 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupCopyJob.BackupDistCp.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupCopyJob.BackupDistCp.html
@@ -36,333 +36,414 @@
 028import org.apache.commons.logging.Log;
 029import 
org.apache.commons.logging.LogFactory;
 030import 
org.apache.hadoop.conf.Configuration;
-031import org.apache.hadoop.fs.FileSystem;
-032import org.apache.hadoop.fs.Path;
-033import 
org.apache.hadoop.hbase.TableName;
-034import 
org.apache.hadoop.hbase.backup.BackupCopyJob;
-035import 
org.apache.hadoop.hbase.backup.BackupInfo;
-036import 
org.apache.hadoop.hbase.backup.BackupType;
-037import 
org.apache.hadoop.hbase.backup.impl.BackupManager;
-038import 
org.apache.hadoop.hbase.backup.util.BackupUtils;
-039import 
org.apache.hadoop.hbase.snapshot.ExportSnapshot;
-040import 
org.apache.hadoop.mapreduce.Cluster;
-041import 
org.apache.hadoop.mapreduce.Counters;
-042import org.apache.hadoop.mapreduce.Job;
-043import 
org.apache.hadoop.mapreduce.JobID;
-044import org.apache.hadoop.tools.DistCp;
-045import 
org.apache.hadoop.tools.DistCpConstants;
-046import 
org.apache.hadoop.tools.DistCpOptions;
-047import 
org.apache.yetus.audience.InterfaceAudience;
-048import 
org.apache.zookeeper.KeeperException.NoNodeException;
-049
-050/**
-051 * Map-Reduce implementation of {@link 
BackupCopyJob}. Basically, there are 2 types of copy
-052 * operation: one is copying from 
snapshot, which bases on extending ExportSnapshot's function, the
-053 * other is copying for incremental log 
files, which bases on extending DistCp's function.
-054 */
-055@InterfaceAudience.Private
-056public class MapReduceBackupCopyJob 
implements BackupCopyJob {
-057  private static final Log LOG = 
LogFactory.getLog(MapReduceBackupCopyJob.class);
-058
-059  private Configuration conf;
-060
-061  // Accumulated progress within the 
whole backup process for the copy operation
-062  private float progressDone = 0.1f;
-063  private long bytesCopied = 0;
-064  private static float INIT_PROGRESS = 
0.1f;
+031import org.apache.hadoop.fs.FileStatus;
+032import org.apache.hadoop.fs.FileSystem;
+033import org.apache.hadoop.fs.Path;
+034import 
org.apache.hadoop.hbase.TableName;
+035import 
org.apache.hadoop.hbase.backup.BackupCopyJob;
+036import 
org.apache.hadoop.hbase.backup.BackupInfo;
+037import 
org.apache.hadoop.hbase.backup.BackupType;
+038import 
org.apache.hadoop.hbase.backup.impl.BackupManager;
+039import 
org.apache.hadoop.hbase.backup.util.BackupUtils;
+040import 
org.apache.hadoop.hbase.snapshot.ExportSnapshot;
+041import 
org.apache.hadoop.io.SequenceFile;
+042import org.apache.hadoop.io.Text;
+043import 
org.apache.hadoop.mapreduce.Cluster;
+044import 
org.apache.hadoop.mapreduce.Counters;
+045import org.apache.hadoop.mapreduce.Job;
+046import 
org.apache.hadoop.mapreduce.JobID;
+047import 
org.apache.hadoop.tools.CopyListingFileStatus;
+048import org.apache.hadoop.tools.DistCp;
+049import 
org.apache.hadoop.tools.DistCpConstants;
+050import 
org.apache.hadoop.tools.DistCpOptions;
+051import 
org.apache.yetus.audience.InterfaceAudience;
+052import 
org.apache.zookeeper.KeeperException.NoNodeException;
+053
+054/**
+055 * Map-Reduce implementation of {@link 
BackupCopyJob}. Basically, there are 2 types of copy
+056 * operation: one is copying from 
snapshot, which bases on extending ExportSnapshot's function, the
+057 * other is copying for incremental log 
files, which bases on extending DistCp's function.
+058 */
+059@InterfaceAudience.Private
+060public class MapReduceBackupCopyJob 
implements BackupCopyJob {
+061  public static final String 
NUMBER_OF_LEVELS_TO_PRESERVE_KEY = "num.levels.preserve";
+062  private static final Log LOG = 
LogFactory.getLog(MapReduceBackupCopyJob.class);
+063
+064  private Configuration conf;
 065
-066  // The percentage of the current copy 
task within the whole task if multiple time copies are
-067  // needed. The default value is 100%, 
which means only 1 copy task for the whole.
-068  private float subTaskPercntgInWholeTask 
= 1f;
-069
-070  public MapReduceBackupCopyJob() {
-071  }
-072
-073  @Override
-074  public Configuration getConf() {
-075return conf;
+066  // Accumulated progress within the 
whole backup process for the copy operation
+067  private float progressDone = 0.1f;
+068  private long bytesCopied = 0;
+069  private static float INIT_PROGRESS = 
0.1f;
+070
+071  // The percentage of the current copy 
task within 

[48/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/book.html
--
diff --git a/book.html b/book.html
index a9efd36..e75c297 100644
--- a/book.html
+++ b/book.html
@@ -35225,7 +35225,7 @@ The server will return cellblocks compressed using this 
same compressor as long
 
 
 Version 3.0.0-SNAPSHOT
-Last updated 2017-10-13 14:29:37 UTC
+Last updated 2017-10-14 14:29:38 UTC
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/bulk-loads.html
--
diff --git a/bulk-loads.html b/bulk-loads.html
index baa09dd..0d572b0 100644
--- a/bulk-loads.html
+++ b/bulk-loads.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase –  
   Bulk Loads in Apache HBase (TM)
@@ -311,7 +311,7 @@ under the License. -->
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-10-13
+  Last Published: 
2017-10-14
 
 
 



[12/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html
--
diff --git 
a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html
 
b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html
index f8309f2..7f75adf 100644
--- 
a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html
+++ 
b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-convergence.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Exemplar for hbase-shaded-client archetype – 
Reactor Dependency Convergence
 
@@ -123,10 +123,10 @@
 305
 
 Number of unique artifacts (NOA):
-331
+329
 
 Number of version-conflicting artifacts (NOC):
-17
+16
 
 Number of SNAPSHOT artifacts (NOS):
 0
@@ -191,75 +191,20 @@
 11.0.2
 
 
-org.apache.hbase:hbase-assembly:pom:3.0.0-SNAPSHOT+- org.apache.hbase:hbase-server:jar:3.0.0-SNAPSHOT:compile|  \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)\- org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-common:jar:2.7.1:compile|  \- (com.google.guava:guava:jar:11.0.2:compile - omitted for 
duplicate)\- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-client-project:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-client:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-endpoint:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-examples:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-mapreduce-client-core:jar:2.7.1:compile|  \- org.apache.hadoop:hadoop-yarn-common:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-hadoop-compat:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-metrics-api:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile  \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-hadoop2-compat:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-it:jar:3.0.0-SNAPSHOT+- org.apache.hbase:hbase-mapreduce:jar:3.0.0-SNAPSHOT:compile|  \- org.apache.hadoop:hadoop-hdfs:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)+- org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT:compile|  \- com.google.guava:guava:jar:11.0.2:compile\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- (com.google.guava:guava:jar:11.0.2:compile - 
omitted for duplicate)
-org.apache.hbase:hbase-mapreduce:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-common:jar:2.7.1:compile|  \- com.google.guava:guava:jar:11.0.2:compile+- org.apache.hadoop:hadoop-hdfs:jar:2.7.1:compile|  \- (com.google.guava:guava:jar:11.0.2:compile - omitted for 
duplicate)\- org.apache.hadoop:hadoop-minicluster:jar:2.7.1:test   +- org.apache.hadoop:hadoop-common:test-jar:tests:2.7.1:test   |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)   \- org.apache.hadoop:hadoop-yarn-server-tests:test-jar:tests:2.7.1:test  +- org.apache.hadoop:hadoop-yarn-server-nodemanager:jar:2.7.1:test  |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  +- org
 .apache.hadoop:hadoop-yarn-server-resourcemanager:jar:2.7.1:test  |  +- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  |  +- org.apache.hadoop:hadoop-yarn-server-applicationhistoryservice:jar:2.7.1:test  |  |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  |  \- org.apache.hadoop:hadoop-yarn-server-web-proxy:jar:2.7.1:test  | \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)
-org.apache.hbase:hbase-metrics-api:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.g

[29/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.Operator.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.Operator.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.Operator.html
index e48f3fd..fcfe8ab 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.Operator.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.Operator.html
@@ -29,536 +29,554 @@
 021import java.io.IOException;
 022import java.util.ArrayList;
 023import java.util.Arrays;
-024import java.util.Collections;
-025import java.util.List;
-026
-027import org.apache.hadoop.hbase.Cell;
-028import 
org.apache.hadoop.hbase.CellComparator;
-029import 
org.apache.hadoop.hbase.CellUtil;
-030import 
org.apache.yetus.audience.InterfaceAudience;
-031import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-032import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-033import 
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
-034import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
-035
-036/**
-037 * Implementation of {@link Filter} that 
represents an ordered List of Filters
-038 * which will be evaluated with a 
specified boolean operator {@link Operator#MUST_PASS_ALL}
-039 * (AND) or 
{@link Operator#MUST_PASS_ONE} (OR).
-040 * Since you can use Filter Lists as 
children of Filter Lists, you can create a
-041 * hierarchy of filters to be 
evaluated.
-042 *
-043 * 
-044 * {@link Operator#MUST_PASS_ALL} evaluates lazily: evaluation stops as soon as one filter does -045 * not include the KeyValue. -046 * -047 *
-048 * {@link Operator#MUST_PASS_ONE} evaluates non-lazily: all filters are always evaluated. -049 * -050 *
-051 * Defaults to {@link Operator#MUST_PASS_ALL}. -052 */ -053@InterfaceAudience.Public -054final public class FilterList extends FilterBase { -055 /** set operator */ -056 @InterfaceAudience.Public -057 public static enum Operator { -058/** !AND */ -059MUST_PASS_ALL, -060/** !OR */ -061MUST_PASS_ONE -062 } -063 -064 private static final int MAX_LOG_FILTERS = 5; -065 private Operator operator = Operator.MUST_PASS_ALL; -066 private final List filters; -067 private Filter seekHintFilter = null; -068 -069 /** Reference Cell used by {@link #transformCell(Cell)} for validation purpose. */ -070 private Cell referenceCell = null; -071 -072 /** -073 * When filtering a given Cell in {@link #filterKeyValue(Cell)}, -074 * this stores the transformed Cell to be returned by {@link #transformCell(Cell)}. -075 * -076 * Individual filters transformation are applied only when the filter includes the Cell. -077 * Transformations are composed in the order specified by {@link #filters}. -078 */ -079 private Cell transformedCell = null; -080 -081 /** -082 * Constructor that takes a set of {@link Filter}s. The default operator -083 * MUST_PASS_ALL is assumed. -084 * All filters are cloned to internal list. -085 * @param rowFilters list of filters -086 */ -087 public FilterList(final List rowFilters) { -088reversed = getReversed(rowFilters, reversed); -089this.filters = new ArrayList<>(rowFilters); -090 } -091 -092 /** -093 * Constructor that takes a var arg number of {@link Filter}s. The fefault operator -094 * MUST_PASS_ALL is assumed. -095 * @param rowFilters -096 */ -097 public FilterList(final Filter... rowFilters) { -098this(Arrays.asList(rowFilters)); -099 } -100 -101 /** -102 * Constructor that takes an operator. -103 * -104 * @param operator Operator to process filter set with. -105 */ -106 public FilterList(final Operator operator) { -107this.operator = operator; -108this.filters = new ArrayList<>(); -109 } -110 -111 /** -112 * Constructor that takes a set of {@link Filter}s and an operator. -113 * -114 * @param operator Operator to process filter set with. -115 * @param rowFilters Set of row filters. -116 */ -117 public FilterList(final Operator operator, final List rowFilters) { -118this(rowFilters); -119this.operator = operator; -120 } -121 -122 /** -123 * Constructor that takes a var arg number of {@link Filter}s and an operator. -124 * -125 * @param operator Operator to process filter set with. -126 * @param rowFilters Filters to use -127 */ -128 public FilterList(final Operator operator, final Filter... rowFilters) { -129this(rowFilters); -130this.operator = operator; -131 } -132 -133 /** -134 * Get the operator. -135 * -136 * @return operator -137 */ -138 public Operator getOperator() { -139return operator; -140 } -141 -142 /** -143 * Get the filters. -144 * -145 * @return filters -146

[47/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/checkstyle-aggregate.html
--
diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html
index 70a127f..a46fc76 100644
--- a/checkstyle-aggregate.html
+++ b/checkstyle-aggregate.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Checkstyle Results
 
@@ -286,10 +286,10 @@
  Warnings
  Errors
 
-2054
+2055
 0
 0
-13589
+13596
 
 Files
 
@@ -797,7 +797,7 @@
 org/apache/hadoop/hbase/backup/impl/IncrementalTableBackupClient.java
 0
 0
-6
+10
 
 org/apache/hadoop/hbase/backup/impl/RestoreTablesClient.java
 0
@@ -2187,7 +2187,7 @@
 org/apache/hadoop/hbase/filter/FilterList.java
 0
 0
-38
+41
 
 org/apache/hadoop/hbase/filter/FilterWrapper.java
 0
@@ -5032,7 +5032,7 @@
 org/apache/hadoop/hbase/regionserver/HRegion.java
 0
 0
-227
+228
 
 org/apache/hadoop/hbase/regionserver/HRegionFileSystem.java
 0
@@ -5102,7 +5102,7 @@
 org/apache/hadoop/hbase/regionserver/KeyValueHeap.java
 0
 0
-12
+11
 
 org/apache/hadoop/hbase/regionserver/KeyValueScanner.java
 0
@@ -8194,12 +8194,12 @@
 
 
 http://checkstyle.sourceforge.net/config_blocks.html#LeftCurly";>LeftCurly
-233
+234
  Error
 
 
 http://checkstyle.sourceforge.net/config_blocks.html#NeedBraces";>NeedBraces
-1693
+1692
  Error
 
 coding
@@ -8261,7 +8261,7 @@
 ordered: "true"
 sortStaticImportsAlphabetically: "true"
 option: "top"
-1830
+1829
  Error
 
 
@@ -8284,19 +8284,19 @@
 caseIndent: "2"
 basicOffset: "2"
 lineWrappingIndentation: "2"
-3694
+3701
  Error
 
 javadoc
 http://checkstyle.sourceforge.net/config_javadoc.html#JavadocTagContinuationIndentation";>JavadocTagContinuationIndentation
 
 offset: "2"
-763
+769
  Error
 
 
 http://checkstyle.sourceforge.net/config_javadoc.html#NonEmptyAtclauseDescription";>NonEmptyAtclauseDescription
-3272
+3265
  Error
 
 misc
@@ -8314,7 +8314,7 @@
 
 max: "100"
 ignorePattern: "^package.*|^import.*|a 
href|href|http://|https://|ftp://|org.apache.thrift.|com.google.protobuf.|hbase.protobuf.generated"
-1096
+1098
  Error
 
 
@@ -14680,7 +14680,7 @@
 
  Error
 javadoc
-NonEmptyAtclauseDescription
+JavadocTagContinuationIndentation
 Javadoc comment at column 43 has parse error. Missed HTML close tag 
'TableName'. Sometimes it means that close tag missed for one of previous 
tags.
 179
 
@@ -15402,40 +15402,64 @@
 Line
 
  Error
-imports
-ImportOrder
-Wrong order for 'org.apache.hadoop.hbase.client.Admin' import.
-47
-
- Error
 blocks
 NeedBraces
 'if' construct must use '{}'s.
-98
-
+96
+
  Error
 sizes
 LineLength
 Line is longer than 100 characters (found 101).
-113
-
+111
+
  Error
 indentation
 Indentation
 'backupManager' have incorrect indentation level 4, expected level should 
be 6.
 116
+
+ Error
+indentation
+Indentation
+'if' have incorrect indentation level 16, expected level should be 14.
+178
 
  Error
-blocks
-NeedBraces
-'if' construct must use '{}'s.
-188
+indentation
+Indentation
+'if' child have incorrect indentation level 18, expected level should be 
16.
+179
+
+ Error
+indentation
+Indentation
+'if rcurly' have incorrect indentation level 16, expected level should be 
14.
+180
+
+ Error
+indentation
+Indentation
+'if' child have incorrect indentation level 16, expected level should be 
14.
+181
 
  Error
+indentation
+Indentation
+'if' child have incorrect indentation level 16, expected level should be 
14.
+184
+
+ Error
 blocks
-NeedBraces
-'if' construct must use '{}'s.
-196
+LeftCurly
+'{' at column 3 should be on the previous line.
+201
+
+ Error
+sizes
+LineLength
+Line is longer than 100 characters (found 105).
+249
 
 org/apache/hadoop/hbase/backup/impl/RestoreTablesClient.java
 
@@ -19081,7 +19105,7 @@
 
  Error
 javadoc
-NonEmptyAtclauseDescription
+JavadocTagContinuationIndentation
 Javadoc comment at column 64 has parse error. Missed HTML close tag 
'code'. Sometimes it means that close tag missed for one of previous tags.
 2103
 
@@ -26011,7 +26035,7 @@
 
  Error
 javadoc
-NonEmptyAtclauseDescription
+JavadocTagContinuationIndentation
 Javadoc comment at column 37 has parse error. Details: no viable 
alternative at input '

[05/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/testdevapidocs/org/apache/hadoop/hbase/coprocessor/example/TestZooKeeperScanPolicyObserver.html
--
diff --git 
a/testdevapidocs/org/apache/hadoop/hbase/coprocessor/example/TestZooKeeperScanPolicyObserver.html
 
b/testdevapidocs/org/apache/hadoop/hbase/coprocessor/example/TestZooKeeperScanPolicyObserver.html
new file mode 100644
index 000..f2c0ace
--- /dev/null
+++ 
b/testdevapidocs/org/apache/hadoop/hbase/coprocessor/example/TestZooKeeperScanPolicyObserver.html
@@ -0,0 +1,489 @@
+http://www.w3.org/TR/html4/loose.dtd";>
+
+
+
+
+
+TestZooKeeperScanPolicyObserver (Apache HBase 3.0.0-SNAPSHOT Test 
API)
+
+
+
+
+
+var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":9,"i5":9,"i6":10};
+var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
+var altColor = "altColor";
+var rowColor = "rowColor";
+var tableTab = "tableTab";
+var activeTableTab = "activeTableTab";
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev Class
+Next Class
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+Summary: 
+Nested | 
+Field | 
+Constr | 
+Method
+
+
+Detail: 
+Field | 
+Constr | 
+Method
+
+
+
+
+
+
+
+
+org.apache.hadoop.hbase.coprocessor.example
+Class 
TestZooKeeperScanPolicyObserver
+
+
+
+http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">java.lang.Object
+
+
+org.apache.hadoop.hbase.coprocessor.example.TestZooKeeperScanPolicyObserver
+
+
+
+
+
+
+
+
+public class TestZooKeeperScanPolicyObserver
+extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
+
+
+
+
+
+
+
+
+
+
+
+Field Summary
+
+Fields 
+
+Modifier and Type
+Field and Description
+
+
+private static byte[]
+FAMILY 
+
+
+private static 
org.apache.hadoop.hbase.TableName
+NAME 
+
+
+private static byte[]
+QUALIFIER 
+
+
+private static 
org.apache.hadoop.hbase.client.Table
+TABLE 
+
+
+private static HBaseTestingUtility
+UTIL 
+
+
+
+
+
+
+
+
+
+Constructor Summary
+
+Constructors 
+
+Constructor and Description
+
+
+TestZooKeeperScanPolicyObserver() 
+
+
+
+
+
+
+
+
+
+Method Summary
+
+All Methods Static Methods Instance Methods Concrete Methods 
+
+Modifier and Type
+Method and Description
+
+
+private void
+assertNotExists(int start,
+   int end) 
+
+
+private void
+assertValueEquals(int start,
+ int end) 
+
+
+private void
+put(int start,
+   int end,
+   long ts) 
+
+
+private void
+setExpireBefore(long time) 
+
+
+static void
+setUp() 
+
+
+static void
+tearDown() 
+
+
+void
+test() 
+
+
+
+
+
+
+Methods inherited from class java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
+http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--";
 title="class or interface in java.lang">clone, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-";
 title="class or interface in java.lang">equals, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--";
 title="class or interface in java.lang">finalize, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--";
 title="class or interface in java.lang">getClass, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--";
 title="class or interface in java.lang">hashCode, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--";
 title="class or interface in java.lang">notify, http://docs.oracle.com/javase/8/docs/api/java/lang
 /Object.html?is-external=true#notifyAll--" title="class or interface in 
java.lang">notifyAll, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--";
 title="class or interface in java.lang">toString, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait--";
 title="class or interface in java.lang">wait, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-";
 title="class or interface in java.lang">wait, http://docs.oracle.com/javase/8/docs/api/java/lang/Ob

[14/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html
--
diff --git 
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html
 
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html
index 55adc09..62e99e5 100644
--- 
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html
+++ 
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-convergence.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Exemplar for hbase-client archetype – Reactor 
Dependency Convergence
 
@@ -123,10 +123,10 @@
 305
 
 Number of unique artifacts (NOA):
-331
+329
 
 Number of version-conflicting artifacts (NOC):
-17
+16
 
 Number of SNAPSHOT artifacts (NOS):
 0
@@ -191,75 +191,20 @@
 11.0.2
 
 
-org.apache.hbase:hbase-assembly:pom:3.0.0-SNAPSHOT+- org.apache.hbase:hbase-server:jar:3.0.0-SNAPSHOT:compile|  \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)\- org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-common:jar:2.7.1:compile|  \- (com.google.guava:guava:jar:11.0.2:compile - omitted for 
duplicate)\- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-client-project:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-client:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-endpoint:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-examples:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-mapreduce-client-core:jar:2.7.1:compile|  \- org.apache.hadoop:hadoop-yarn-common:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-hadoop-compat:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-metrics-api:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile  \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-hadoop2-compat:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-it:jar:3.0.0-SNAPSHOT+- org.apache.hbase:hbase-mapreduce:jar:3.0.0-SNAPSHOT:compile|  \- org.apache.hadoop:hadoop-hdfs:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)+- org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT:compile|  \- com.google.guava:guava:jar:11.0.2:compile\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- (com.google.guava:guava:jar:11.0.2:compile - 
omitted for duplicate)
-org.apache.hbase:hbase-mapreduce:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-common:jar:2.7.1:compile|  \- com.google.guava:guava:jar:11.0.2:compile+- org.apache.hadoop:hadoop-hdfs:jar:2.7.1:compile|  \- (com.google.guava:guava:jar:11.0.2:compile - omitted for 
duplicate)\- org.apache.hadoop:hadoop-minicluster:jar:2.7.1:test   +- org.apache.hadoop:hadoop-common:test-jar:tests:2.7.1:test   |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)   \- org.apache.hadoop:hadoop-yarn-server-tests:test-jar:tests:2.7.1:test  +- org.apache.hadoop:hadoop-yarn-server-nodemanager:jar:2.7.1:test  |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  +- org
 .apache.hadoop:hadoop-yarn-server-resourcemanager:jar:2.7.1:test  |  +- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  |  +- org.apache.hadoop:hadoop-yarn-server-applicationhistoryservice:jar:2.7.1:test  |  |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  |  \- org.apache.hadoop:hadoop-yarn-server-web-proxy:jar:2.7.1:test  | \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)
-org.apache.hbase:hbase-metrics-api:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.

[16/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-build-configuration/hbase-archetypes/hbase-archetype-builder/dependency-convergence.html
--
diff --git 
a/hbase-build-configuration/hbase-archetypes/hbase-archetype-builder/dependency-convergence.html
 
b/hbase-build-configuration/hbase-archetypes/hbase-archetype-builder/dependency-convergence.html
index 2694357..3d08ece 100644
--- 
a/hbase-build-configuration/hbase-archetypes/hbase-archetype-builder/dependency-convergence.html
+++ 
b/hbase-build-configuration/hbase-archetypes/hbase-archetype-builder/dependency-convergence.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Archetype builder – Reactor Dependency 
Convergence
 
@@ -123,10 +123,10 @@
 305
 
 Number of unique artifacts (NOA):
-331
+329
 
 Number of version-conflicting artifacts (NOC):
-17
+16
 
 Number of SNAPSHOT artifacts (NOS):
 0
@@ -191,75 +191,20 @@
 11.0.2
 
 
-org.apache.hbase:hbase-assembly:pom:3.0.0-SNAPSHOT+- org.apache.hbase:hbase-server:jar:3.0.0-SNAPSHOT:compile|  \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)\- org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-common:jar:2.7.1:compile|  \- (com.google.guava:guava:jar:11.0.2:compile - omitted for 
duplicate)\- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-client-project:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-client:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-endpoint:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-examples:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-mapreduce-client-core:jar:2.7.1:compile|  \- org.apache.hadoop:hadoop-yarn-common:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-hadoop-compat:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-metrics-api:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile  \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-hadoop2-compat:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-it:jar:3.0.0-SNAPSHOT+- org.apache.hbase:hbase-mapreduce:jar:3.0.0-SNAPSHOT:compile|  \- org.apache.hadoop:hadoop-hdfs:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)+- org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT:compile|  \- com.google.guava:guava:jar:11.0.2:compile\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- (com.google.guava:guava:jar:11.0.2:compile - 
omitted for duplicate)
-org.apache.hbase:hbase-mapreduce:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-common:jar:2.7.1:compile|  \- com.google.guava:guava:jar:11.0.2:compile+- org.apache.hadoop:hadoop-hdfs:jar:2.7.1:compile|  \- (com.google.guava:guava:jar:11.0.2:compile - omitted for 
duplicate)\- org.apache.hadoop:hadoop-minicluster:jar:2.7.1:test   +- org.apache.hadoop:hadoop-common:test-jar:tests:2.7.1:test   |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)   \- org.apache.hadoop:hadoop-yarn-server-tests:test-jar:tests:2.7.1:test  +- org.apache.hadoop:hadoop-yarn-server-nodemanager:jar:2.7.1:test  |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  +- org
 .apache.hadoop:hadoop-yarn-server-resourcemanager:jar:2.7.1:test  |  +- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  |  +- org.apache.hadoop:hadoop-yarn-server-applicationhistoryservice:jar:2.7.1:test  |  |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  |  \- org.apache.hadoop:hadoop-yarn-server-web-proxy:jar:2.7.1:test  | \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)
-org.apache.hbase:hbase-metrics-api:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hba

[22/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-annotations/dependency-convergence.html
--
diff --git a/hbase-annotations/dependency-convergence.html 
b/hbase-annotations/dependency-convergence.html
index 89abfd2..6db01ca 100644
--- a/hbase-annotations/dependency-convergence.html
+++ b/hbase-annotations/dependency-convergence.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Annotations – Reactor Dependency 
Convergence
 
@@ -123,10 +123,10 @@
 305
 
 Number of unique artifacts (NOA):
-331
+329
 
 Number of version-conflicting artifacts (NOC):
-17
+16
 
 Number of SNAPSHOT artifacts (NOS):
 0
@@ -191,75 +191,20 @@
 11.0.2
 
 
-org.apache.hbase:hbase-assembly:pom:3.0.0-SNAPSHOT+- org.apache.hbase:hbase-server:jar:3.0.0-SNAPSHOT:compile|  \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)\- org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-common:jar:2.7.1:compile|  \- (com.google.guava:guava:jar:11.0.2:compile - omitted for 
duplicate)\- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-client-project:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-client:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-endpoint:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-examples:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-mapreduce-client-core:jar:2.7.1:compile|  \- org.apache.hadoop:hadoop-yarn-common:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-hadoop-compat:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-metrics-api:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile  \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-hadoop2-compat:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-it:jar:3.0.0-SNAPSHOT+- org.apache.hbase:hbase-mapreduce:jar:3.0.0-SNAPSHOT:compile|  \- org.apache.hadoop:hadoop-hdfs:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)+- org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT:compile|  \- com.google.guava:guava:jar:11.0.2:compile\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- (com.google.guava:guava:jar:11.0.2:compile - 
omitted for duplicate)
-org.apache.hbase:hbase-mapreduce:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-common:jar:2.7.1:compile|  \- com.google.guava:guava:jar:11.0.2:compile+- org.apache.hadoop:hadoop-hdfs:jar:2.7.1:compile|  \- (com.google.guava:guava:jar:11.0.2:compile - omitted for 
duplicate)\- org.apache.hadoop:hadoop-minicluster:jar:2.7.1:test   +- org.apache.hadoop:hadoop-common:test-jar:tests:2.7.1:test   |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)   \- org.apache.hadoop:hadoop-yarn-server-tests:test-jar:tests:2.7.1:test  +- org.apache.hadoop:hadoop-yarn-server-nodemanager:jar:2.7.1:test  |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  +- org
 .apache.hadoop:hadoop-yarn-server-resourcemanager:jar:2.7.1:test  |  +- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  |  +- org.apache.hadoop:hadoop-yarn-server-applicationhistoryservice:jar:2.7.1:test  |  |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  |  \- org.apache.hadoop:hadoop-yarn-server-web-proxy:jar:2.7.1:test  | \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)
-org.apache.hbase:hbase-metrics-api:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-metrics:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-procedure:jar:3.0.0-SNAPSHOT\- 

[35/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/ScannerContext.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/ScannerContext.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/ScannerContext.html
index 73706c1..2daa6c1 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/ScannerContext.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/ScannerContext.html
@@ -83,6 +83,10 @@
 
 
 
+org.apache.hadoop.hbase.coprocessor.example
+ 
+
+
 org.apache.hadoop.hbase.regionserver
  
 
@@ -91,6 +95,25 @@
 
 
 
+
+
+
+Uses of ScannerContext in org.apache.hadoop.hbase.coprocessor.example
+
+Methods in org.apache.hadoop.hbase.coprocessor.example
 with parameters of type ScannerContext 
+
+Modifier and Type
+Method and Description
+
+
+
+boolean
+DelegatingInternalScanner.next(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List result,
+ScannerContext scannerContext) 
+
+
+
+
 
 
 
@@ -199,7 +222,9 @@
 
 boolean
 KeyValueHeap.next(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List result,
-ScannerContext scannerContext) 
+ScannerContext scannerContext)
+Gets the next row of keys from the top-most scanner.
+
 
 
 boolean

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/Store.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/Store.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/Store.html
index 2238d57..abbc7bf 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/Store.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/class-use/Store.html
@@ -231,38 +231,19 @@
 
 
 
-protected ScanInfo
-ZooKeeperScanPolicyObserver.getScanInfo(Store store,
-   RegionCoprocessorEnvironment e) 
-
-
 InternalScanner
-ZooKeeperScanPolicyObserver.preCompactScannerOpen(ObserverContext
  c,
- Store store,
- http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List scanners,
- ScanType scanType,
- long earliestPutTs,
- InternalScanner s,
- CompactionLifeCycleTracker tracker,
- CompactionRequest request,
- long readPoint) 
-
-
-InternalScanner
-ZooKeeperScanPolicyObserver.preFlushScannerOpen(ObserverContext c,
-   Store store,
-   http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List scanners,
-   InternalScanner s,
-   long readPoint) 
+ZooKeeperScanPolicyObserver.preCompact(ObserverContext c,
+  Store store,
+  InternalScanner scanner,
+  ScanType scanType,
+  CompactionLifeCycleTracker tracker,
+  CompactionRequest request) 
 
 
-KeyValueScanner
-ZooKeeperScanPolicyObserver.preStoreScannerOpen(ObserverContext c,
-   Store store,
-   Scan scan,
-   http://docs.oracle.com/javase/8/docs/api/java/util/NavigableSet.html?is-external=true";
 title="class or interface in 
java.util">NavigableSet targetCols,
-   KeyValueScanner s,
-   long readPoint) 
+InternalScanner
+ZooKeeperScanPolicyObserver.preFlush(ObserverContext c,
+Store store,
+InternalScanner scanner) 
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/org/apache/hadoop/hbase/regionserver/compactions/class-use/CompactionLifeCycleTracker.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/regionserver/compactions/class-use/CompactionLifeCycleTracker.html
 
b/devapidocs/org/apache/hadoop/hbase/regionserver/compactions/class-use/CompactionLifeCycleTracker.html
index d9f6f93..36e3daf 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/regionserver/compactions/class-use/CompactionLifeCycleTracker.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/regionserver/compactions/class-use/CompactionLifeCycleTracker.html
@@ -194,15 +194,12 @@
 
 
 InternalScanner
-ZooKeeperScanPolicyObserver.preCompactScannerOpen(ObserverContext c,
- Store store,
- 

[38/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/org/apache/hadoop/hbase/coprocessor/example/class-use/DelegatingInternalScanner.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/class-use/DelegatingInternalScanner.html
 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/class-use/DelegatingInternalScanner.html
new file mode 100644
index 000..b820e37
--- /dev/null
+++ 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/class-use/DelegatingInternalScanner.html
@@ -0,0 +1,125 @@
+http://www.w3.org/TR/html4/loose.dtd";>
+
+
+
+
+
+Uses of Class 
org.apache.hadoop.hbase.coprocessor.example.DelegatingInternalScanner (Apache 
HBase 3.0.0-SNAPSHOT API)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev
+Next
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+
+
+
+Uses of 
Classorg.apache.hadoop.hbase.coprocessor.example.DelegatingInternalScanner
+
+No usage of 
org.apache.hadoop.hbase.coprocessor.example.DelegatingInternalScanner
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev
+Next
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+
+
+Copyright © 2007–2017 https://www.apache.org/";>The Apache Software Foundation. All rights 
reserved.
+
+

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/org/apache/hadoop/hbase/coprocessor/example/class-use/ZooKeeperScanPolicyObserver.ZKDataHolder.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/class-use/ZooKeeperScanPolicyObserver.ZKDataHolder.html
 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/class-use/ZooKeeperScanPolicyObserver.ZKDataHolder.html
new file mode 100644
index 000..cee2724
--- /dev/null
+++ 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/class-use/ZooKeeperScanPolicyObserver.ZKDataHolder.html
@@ -0,0 +1,125 @@
+http://www.w3.org/TR/html4/loose.dtd";>
+
+
+
+
+
+Uses of Class 
org.apache.hadoop.hbase.coprocessor.example.ZooKeeperScanPolicyObserver.ZKDataHolder
 (Apache HBase 3.0.0-SNAPSHOT API)
+
+
+
+
+
+
+
+JavaScript is disabled on your browser.
+
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev
+Next
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+
+
+
+Uses of 
Classorg.apache.hadoop.hbase.coprocessor.example.ZooKeeperScanPolicyObserver.ZKDataHolder
+
+No usage of 
org.apache.hadoop.hbase.coprocessor.example.ZooKeeperScanPolicyObserver.ZKDataHolder
+
+
+
+
+Skip navigation links
+
+
+
+
+Overview
+Package
+Class
+Use
+Tree
+Deprecated
+Index
+Help
+
+
+
+
+Prev
+Next
+
+
+Frames
+No Frames
+
+
+All Classes
+
+
+
+
+
+
+
+
+
+Copyright © 2007–2017 https://www.apache.org/";>The Apache Software Foundation. All rights 
reserved.
+
+

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/org/apache/hadoop/hbase/coprocessor/example/class-use/ZooKeeperScanPolicyObserver.ZKWatcher.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/class-use/ZooKeeperScanPolicyObserver.ZKWatcher.html
 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/class-use/ZooKeeperScanPolicyObserver.ZKWatcher

[39/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.html
 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.html
index 2553718..2a380e3 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.html
@@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab";
 
 
 Prev Class
-Next Class
+Next Class
 
 
 Frames
@@ -113,24 +113,17 @@ var activeTableTab = "activeTableTab";
 
 
 
-public class ZooKeeperScanPolicyObserver
+public class ZooKeeperScanPolicyObserver
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
 implements RegionCoprocessor, RegionObserver
-This is an example showing how a RegionObserver could 
configured
- via ZooKeeper in order to control a Region compaction, flush, and scan policy.
-
- This also demonstrated the use of shared
- RegionObserver state.
- See RegionCoprocessorEnvironment.getSharedData().
-
- This would be useful for an incremental backup tool, which would indicate the 
last
- time of a successful backup via ZK and instruct HBase to not delete data that 
was
- inserted since (based on wall clock time).
-
- This implements org.apache.zookeeper.Watcher directly instead of using
- ZooKeeperWatcher,
- because RegionObservers come and go and currently
- listeners registered with ZooKeeperWatcher cannot be removed.
+This is an example showing how a RegionObserver could 
configured via ZooKeeper in order to
+ control a Region compaction, flush, and scan policy. This also demonstrated 
the use of shared
+ RegionObserver state. See
+ RegionCoprocessorEnvironment.getSharedData().
+ 
+ This would be useful for an incremental backup tool, which would indicate the 
last time of a
+ successful backup via ZK and instruct HBase that to safely delete the data 
which has already been
+ backup.
 
 
 
@@ -151,7 +144,7 @@ implements 
 private static class 
-ZooKeeperScanPolicyObserver.ZKWatcher
+ZooKeeperScanPolicyObserver.ZKDataHolder
 Internal watcher that keep "data" up to date 
asynchronously.
 
 
@@ -185,32 +178,28 @@ implements Field and Description
 
 
-private static 
org.apache.commons.logging.Log
-LOG 
+private 
org.apache.curator.framework.recipes.cache.NodeCache
+cache 
 
 
 static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
-node 
+NODE 
 
 
-private org.apache.zookeeper.ZooKeeper
-zk 
-
-
 static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
 ZK_ENSEMBLE_KEY 
 
-
+
 static int
 ZK_SESSION_TIMEOUT_DEFAULT 
 
-
+
 static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
 ZK_SESSION_TIMEOUT_KEY 
 
-
-static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
-zkkey 
+
+private static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
+ZKKEY 
 
 
 
@@ -252,65 +241,51 @@ implements Method and Description
 
 
-http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true";
 title="class or interface in java.util">Optional
-getRegionObserver() 
+private http://docs.oracle.com/javase/8/docs/api/java/util/OptionalLong.html?is-external=true";
 title="class or interface in java.util">OptionalLong
+getExpireBefore() 
 
 
-protected ScanInfo
-getScanInfo(Store store,
-   RegionCoprocessorEnvironment e) 
+http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true";
 title="class or interface in java.util">Optional
+getRegionObserver() 
 
 
 InternalScanner
-preCompactScannerOpen(ObserverContext c,
- Store store,
- http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List scanners,
- ScanType scanType,
- long earliestPutTs,
- InternalScanner s,
- CompactionLifeCycleTracker tracker,
- CompactionRequest request,
- long readPoint)
+preCompact(ObserverContext c,
+  Store store,
+  InternalScanner scanner,
+  ScanType scanType,
+  CompactionLifeCycleTracker tracker,
+

[31/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupCopyJob.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupCopyJob.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupCopyJob.html
index 65a31ce..7551292 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupCopyJob.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupCopyJob.html
@@ -36,333 +36,414 @@
 028import org.apache.commons.logging.Log;
 029import 
org.apache.commons.logging.LogFactory;
 030import 
org.apache.hadoop.conf.Configuration;
-031import org.apache.hadoop.fs.FileSystem;
-032import org.apache.hadoop.fs.Path;
-033import 
org.apache.hadoop.hbase.TableName;
-034import 
org.apache.hadoop.hbase.backup.BackupCopyJob;
-035import 
org.apache.hadoop.hbase.backup.BackupInfo;
-036import 
org.apache.hadoop.hbase.backup.BackupType;
-037import 
org.apache.hadoop.hbase.backup.impl.BackupManager;
-038import 
org.apache.hadoop.hbase.backup.util.BackupUtils;
-039import 
org.apache.hadoop.hbase.snapshot.ExportSnapshot;
-040import 
org.apache.hadoop.mapreduce.Cluster;
-041import 
org.apache.hadoop.mapreduce.Counters;
-042import org.apache.hadoop.mapreduce.Job;
-043import 
org.apache.hadoop.mapreduce.JobID;
-044import org.apache.hadoop.tools.DistCp;
-045import 
org.apache.hadoop.tools.DistCpConstants;
-046import 
org.apache.hadoop.tools.DistCpOptions;
-047import 
org.apache.yetus.audience.InterfaceAudience;
-048import 
org.apache.zookeeper.KeeperException.NoNodeException;
-049
-050/**
-051 * Map-Reduce implementation of {@link 
BackupCopyJob}. Basically, there are 2 types of copy
-052 * operation: one is copying from 
snapshot, which bases on extending ExportSnapshot's function, the
-053 * other is copying for incremental log 
files, which bases on extending DistCp's function.
-054 */
-055@InterfaceAudience.Private
-056public class MapReduceBackupCopyJob 
implements BackupCopyJob {
-057  private static final Log LOG = 
LogFactory.getLog(MapReduceBackupCopyJob.class);
-058
-059  private Configuration conf;
-060
-061  // Accumulated progress within the 
whole backup process for the copy operation
-062  private float progressDone = 0.1f;
-063  private long bytesCopied = 0;
-064  private static float INIT_PROGRESS = 
0.1f;
+031import org.apache.hadoop.fs.FileStatus;
+032import org.apache.hadoop.fs.FileSystem;
+033import org.apache.hadoop.fs.Path;
+034import 
org.apache.hadoop.hbase.TableName;
+035import 
org.apache.hadoop.hbase.backup.BackupCopyJob;
+036import 
org.apache.hadoop.hbase.backup.BackupInfo;
+037import 
org.apache.hadoop.hbase.backup.BackupType;
+038import 
org.apache.hadoop.hbase.backup.impl.BackupManager;
+039import 
org.apache.hadoop.hbase.backup.util.BackupUtils;
+040import 
org.apache.hadoop.hbase.snapshot.ExportSnapshot;
+041import 
org.apache.hadoop.io.SequenceFile;
+042import org.apache.hadoop.io.Text;
+043import 
org.apache.hadoop.mapreduce.Cluster;
+044import 
org.apache.hadoop.mapreduce.Counters;
+045import org.apache.hadoop.mapreduce.Job;
+046import 
org.apache.hadoop.mapreduce.JobID;
+047import 
org.apache.hadoop.tools.CopyListingFileStatus;
+048import org.apache.hadoop.tools.DistCp;
+049import 
org.apache.hadoop.tools.DistCpConstants;
+050import 
org.apache.hadoop.tools.DistCpOptions;
+051import 
org.apache.yetus.audience.InterfaceAudience;
+052import 
org.apache.zookeeper.KeeperException.NoNodeException;
+053
+054/**
+055 * Map-Reduce implementation of {@link 
BackupCopyJob}. Basically, there are 2 types of copy
+056 * operation: one is copying from 
snapshot, which bases on extending ExportSnapshot's function, the
+057 * other is copying for incremental log 
files, which bases on extending DistCp's function.
+058 */
+059@InterfaceAudience.Private
+060public class MapReduceBackupCopyJob 
implements BackupCopyJob {
+061  public static final String 
NUMBER_OF_LEVELS_TO_PRESERVE_KEY = "num.levels.preserve";
+062  private static final Log LOG = 
LogFactory.getLog(MapReduceBackupCopyJob.class);
+063
+064  private Configuration conf;
 065
-066  // The percentage of the current copy 
task within the whole task if multiple time copies are
-067  // needed. The default value is 100%, 
which means only 1 copy task for the whole.
-068  private float subTaskPercntgInWholeTask 
= 1f;
-069
-070  public MapReduceBackupCopyJob() {
-071  }
-072
-073  @Override
-074  public Configuration getConf() {
-075return conf;
+066  // Accumulated progress within the 
whole backup process for the copy operation
+067  private float progressDone = 0.1f;
+068  private long bytesCopied = 0;
+069  private static float INIT_PROGRESS = 
0.1f;
+070
+071  // The percentage of the current copy 
task within the whole task if multiple time copies are
+072  // needed. The d

[41/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
index 9485a04..c608b83 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/package-tree.html
@@ -549,25 +549,25 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.client.TableState.State
-org.apache.hadoop.hbase.client.AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState
-org.apache.hadoop.hbase.client.AbstractResponse.ResponseType
+org.apache.hadoop.hbase.client.Durability
 org.apache.hadoop.hbase.client.MobCompactPartitionPolicy
-org.apache.hadoop.hbase.client.SnapshotType
 org.apache.hadoop.hbase.client.Consistency
 org.apache.hadoop.hbase.client.AsyncScanSingleRegionRpcRetryingCaller.ScanControllerState
-org.apache.hadoop.hbase.client.RequestController.ReturnCode
-org.apache.hadoop.hbase.client.RegionLocateType
-org.apache.hadoop.hbase.client.CompactionState
-org.apache.hadoop.hbase.client.MasterSwitchType
-org.apache.hadoop.hbase.client.IsolationLevel
-org.apache.hadoop.hbase.client.CompactType
-org.apache.hadoop.hbase.client.Durability
-org.apache.hadoop.hbase.client.AsyncProcessTask.SubmittedRows
-org.apache.hadoop.hbase.client.ScannerCallable.MoreResults
 org.apache.hadoop.hbase.client.AsyncRequestFutureImpl.Retry
+org.apache.hadoop.hbase.client.AbstractResponse.ResponseType
+org.apache.hadoop.hbase.client.ScannerCallable.MoreResults
+org.apache.hadoop.hbase.client.CompactType
 org.apache.hadoop.hbase.client.Scan.ReadType
+org.apache.hadoop.hbase.client.MasterSwitchType
+org.apache.hadoop.hbase.client.RegionLocateType
+org.apache.hadoop.hbase.client.SnapshotType
+org.apache.hadoop.hbase.client.AsyncProcessTask.SubmittedRows
 org.apache.hadoop.hbase.client.HBaseAdmin.ReplicationState
+org.apache.hadoop.hbase.client.RequestController.ReturnCode
+org.apache.hadoop.hbase.client.CompactionState
+org.apache.hadoop.hbase.client.TableState.State
+org.apache.hadoop.hbase.client.AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState
+org.apache.hadoop.hbase.client.IsolationLevel
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/org/apache/hadoop/hbase/client/package-use.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/client/package-use.html 
b/devapidocs/org/apache/hadoop/hbase/client/package-use.html
index 4277523..a2448b8 100644
--- a/devapidocs/org/apache/hadoop/hbase/client/package-use.html
+++ b/devapidocs/org/apache/hadoop/hbase/client/package-use.html
@@ -1875,11 +1875,6 @@ service.
 
 
 
-Scan
-Used to perform Scan operations.
-
-
-
 TableDescriptor
 TableDescriptor contains the details about an HBase table 
such as the descriptors of
  all the column families, is the table a catalog table,  hbase:meta 
,

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
index ac98e78..fc61ce4 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
@@ -2047,15 +2047,12 @@
 
 
 InternalScanner
-ZooKeeperScanPolicyObserver.preCompactScannerOpen(ObserverContext
  c,
- Store store,
- http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List scanners,
- ScanType scanType,
- long earliestPutTs,
- InternalScanner s,
- CompactionLifeCycleTracker tracker,
- CompactionRequest request,
- long readPoint) 
+ZooKeeperScanPolicyObserver.preCompact(ObserverContext c,
+  Store store,
+  InternalScanner scanner,
+  ScanType scanType,
+  CompactionLifeCycleTracker tracker,
+  CompactionRequest request) 
 
 
 void

[44/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/dependency-info.html
--
diff --git a/dependency-info.html b/dependency-info.html
index a0a3543..a4f7e38 100644
--- a/dependency-info.html
+++ b/dependency-info.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Dependency Information
 
@@ -318,7 +318,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-10-13
+  Last Published: 
2017-10-14
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/dependency-management.html
--
diff --git a/dependency-management.html b/dependency-management.html
index 3f4f8db..2f985dd 100644
--- a/dependency-management.html
+++ b/dependency-management.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Project Dependency Management
 
@@ -413,384 +413,390 @@
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>The Apache Software 
License, Version 2.0
 
+org.apache.curator
+http://curator.apache.org/curator-recipes";>curator-recipes
+4.0.0
+jar
+http://www.apache.org/licenses/LICENSE-2.0.txt";>The Apache Software 
License, Version 2.0
+
 org.apache.hadoop
 hadoop-auth
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-client
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-common
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-hdfs
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-mapreduce-client-core
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-mapreduce-client-jobclient
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-minicluster
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-annotations";>hbase-annotations
 3.0.0-SNAPSHOT
 test-jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-client";>hbase-client
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-common";>hbase-common
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-endpoint";>hbase-endpoint
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-support/hbase-error-prone";>hbase-error-prone
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-examples";>hbase-examples
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-external-blockcache";>hbase-external-blockcache
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-hadoop-compat";>hbase-hadoop-compat
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-hadoop2-compat";>hbase-hadoop2-compat
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-mapreduce";>hbase-mapreduce
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-metrics";>hbase-metrics
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-metrics-api";>hbase-metrics-api
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-procedure";>hbase-procedure
 3.0.0-SNAPSHOT
 jar
 ht

[28/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.html
--
diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.html
index e48f3fd..fcfe8ab 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.html
@@ -29,536 +29,554 @@
 021import java.io.IOException;
 022import java.util.ArrayList;
 023import java.util.Arrays;
-024import java.util.Collections;
-025import java.util.List;
-026
-027import org.apache.hadoop.hbase.Cell;
-028import 
org.apache.hadoop.hbase.CellComparator;
-029import 
org.apache.hadoop.hbase.CellUtil;
-030import 
org.apache.yetus.audience.InterfaceAudience;
-031import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-032import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-033import 
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
-034import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
-035
-036/**
-037 * Implementation of {@link Filter} that 
represents an ordered List of Filters
-038 * which will be evaluated with a 
specified boolean operator {@link Operator#MUST_PASS_ALL}
-039 * (AND) or 
{@link Operator#MUST_PASS_ONE} (OR).
-040 * Since you can use Filter Lists as 
children of Filter Lists, you can create a
-041 * hierarchy of filters to be 
evaluated.
-042 *
-043 * 
-044 * {@link Operator#MUST_PASS_ALL} evaluates lazily: evaluation stops as soon as one filter does -045 * not include the KeyValue. -046 * -047 *
-048 * {@link Operator#MUST_PASS_ONE} evaluates non-lazily: all filters are always evaluated. -049 * -050 *
-051 * Defaults to {@link Operator#MUST_PASS_ALL}. -052 */ -053@InterfaceAudience.Public -054final public class FilterList extends FilterBase { -055 /** set operator */ -056 @InterfaceAudience.Public -057 public static enum Operator { -058/** !AND */ -059MUST_PASS_ALL, -060/** !OR */ -061MUST_PASS_ONE -062 } -063 -064 private static final int MAX_LOG_FILTERS = 5; -065 private Operator operator = Operator.MUST_PASS_ALL; -066 private final List filters; -067 private Filter seekHintFilter = null; -068 -069 /** Reference Cell used by {@link #transformCell(Cell)} for validation purpose. */ -070 private Cell referenceCell = null; -071 -072 /** -073 * When filtering a given Cell in {@link #filterKeyValue(Cell)}, -074 * this stores the transformed Cell to be returned by {@link #transformCell(Cell)}. -075 * -076 * Individual filters transformation are applied only when the filter includes the Cell. -077 * Transformations are composed in the order specified by {@link #filters}. -078 */ -079 private Cell transformedCell = null; -080 -081 /** -082 * Constructor that takes a set of {@link Filter}s. The default operator -083 * MUST_PASS_ALL is assumed. -084 * All filters are cloned to internal list. -085 * @param rowFilters list of filters -086 */ -087 public FilterList(final List rowFilters) { -088reversed = getReversed(rowFilters, reversed); -089this.filters = new ArrayList<>(rowFilters); -090 } -091 -092 /** -093 * Constructor that takes a var arg number of {@link Filter}s. The fefault operator -094 * MUST_PASS_ALL is assumed. -095 * @param rowFilters -096 */ -097 public FilterList(final Filter... rowFilters) { -098this(Arrays.asList(rowFilters)); -099 } -100 -101 /** -102 * Constructor that takes an operator. -103 * -104 * @param operator Operator to process filter set with. -105 */ -106 public FilterList(final Operator operator) { -107this.operator = operator; -108this.filters = new ArrayList<>(); -109 } -110 -111 /** -112 * Constructor that takes a set of {@link Filter}s and an operator. -113 * -114 * @param operator Operator to process filter set with. -115 * @param rowFilters Set of row filters. -116 */ -117 public FilterList(final Operator operator, final List rowFilters) { -118this(rowFilters); -119this.operator = operator; -120 } -121 -122 /** -123 * Constructor that takes a var arg number of {@link Filter}s and an operator. -124 * -125 * @param operator Operator to process filter set with. -126 * @param rowFilters Filters to use -127 */ -128 public FilterList(final Operator operator, final Filter... rowFilters) { -129this(rowFilters); -130this.operator = operator; -131 } -132 -133 /** -134 * Get the operator. -135 * -136 * @return operator -137 */ -138 public Operator getOperator() { -139return operator; -140 } -141 -142 /** -143 * Get the filters. -144 * -145 * @return filters -146 */ -147 public List getFilters() { -14

[46/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/checkstyle.rss
--
diff --git a/checkstyle.rss b/checkstyle.rss
index b616326..acd2adb 100644
--- a/checkstyle.rss
+++ b/checkstyle.rss
@@ -25,8 +25,8 @@ under the License.
 en-us
 ©2007 - 2017 The Apache Software Foundation
 
-  File: 2054,
- Errors: 13589,
+  File: 2055,
+ Errors: 13596,
  Warnings: 0,
  Infos: 0
   
@@ -12851,7 +12851,7 @@ under the License.
   0
 
 
-  38
+  41
 
   
   
@@ -20089,7 +20089,7 @@ under the License.
   0
 
 
-  12
+  11
 
   
   
@@ -21419,7 +21419,7 @@ under the License.
   0
 
 
-  6
+  10
 
   
   
@@ -22236,6 +22236,20 @@ under the License.
   
   
 
+  http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.coprocessor.example.DelegatingInternalScanner.java";>org/apache/hadoop/hbase/coprocessor/example/DelegatingInternalScanner.java
+
+
+  0
+
+
+  0
+
+
+  0
+
+  
+  
+
   http://hbase.apache.org/checkstyle.html#org.apache.hadoop.hbase.NoTagsByteBufferKeyValue.java";>org/apache/hadoop/hbase/NoTagsByteBufferKeyValue.java
 
 
@@ -26067,7 +26081,7 @@ under the License.
   0
 
 
-  227
+  228
 
   
   

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/coc.html
--
diff --git a/coc.html b/coc.html
index efc01ef..557274f 100644
--- a/coc.html
+++ b/coc.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – 
   Code of Conduct Policy
@@ -380,7 +380,7 @@ email to mailto:priv...@hbase.apache.org";>the priv
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-10-13
+  Last Published: 
2017-10-14
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/cygwin.html
--
diff --git a/cygwin.html b/cygwin.html
index c9bdb82..82661b2 100644
--- a/cygwin.html
+++ b/cygwin.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Installing Apache HBase (TM) on Windows using 
Cygwin
 
@@ -679,7 +679,7 @@ Now your HBase server is running, start 
coding and build that next
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-10-13
+  Last Published: 
2017-10-14
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/dependencies.html
--
diff --git a/dependencies.html b/dependencies.html
index 15ed21c..0c1d34d 100644
--- a/dependencies.html
+++ b/dependencies.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Project Dependencies
 
@@ -445,7 +445,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-10-13
+  Last Published: 
2017-10-14
 
 
 



[01/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
Repository: hbase-site
Updated Branches:
  refs/heads/asf-site c1db45023 -> 7816cbdea


http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/testdevapidocs/src-html/org/apache/hadoop/hbase/backup/TestBackupBase.html
--
diff --git 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/backup/TestBackupBase.html 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/backup/TestBackupBase.html
index 11a0572..4c6fde8 100644
--- a/testdevapidocs/src-html/org/apache/hadoop/hbase/backup/TestBackupBase.html
+++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/backup/TestBackupBase.html
@@ -144,375 +144,376 @@
 136// copy out the table and region 
info files for each table
 137
BackupUtils.copyTableRegionInfo(conn, backupInfo, conf);
 138// convert WAL to HFiles and copy 
them to .tmp under BACKUP_ROOT
-139
convertWALsToHFiles(backupInfo);
-140
incrementalCopyHFiles(backupInfo);
-141failStageIf(Stage.stage_2);
-142// Save list of WAL files 
copied
-143
backupManager.recordWALFiles(backupInfo.getIncrBackupFileList());
-144
-145// case INCR_BACKUP_COMPLETE:
-146// set overall backup status: 
complete. Here we make sure to complete the backup.
-147// After this checkpoint, even if 
entering cancel process, will let the backup finished
-148// Set the previousTimestampMap 
which is before this current log roll to the manifest.
-149HashMap> previousTimestampMap =
-150
backupManager.readLogTimestampMap();
-151
backupInfo.setIncrTimestampMap(previousTimestampMap);
-152
-153// The table list in backupInfo 
is good for both full backup and incremental backup.
-154// For incremental backup, it 
contains the incremental backup table set.
-155
backupManager.writeRegionServerLogTimestamp(backupInfo.getTables(), 
newTimestamps);
-156failStageIf(Stage.stage_3);
-157
-158HashMap> newTableSetTimestampMap =
-159
backupManager.readLogTimestampMap();
-160
-161Long newStartCode =
-162
BackupUtils.getMinValue(BackupUtils.getRSLogTimestampMins(newTableSetTimestampMap));
-163
backupManager.writeBackupStartCode(newStartCode);
-164
-165
handleBulkLoad(backupInfo.getTableNames());
-166failStageIf(Stage.stage_4);
-167
-168// backup complete
-169completeBackup(conn, backupInfo, 
backupManager, BackupType.INCREMENTAL, conf);
-170
-171  } catch (Exception e) {
-172failBackup(conn, backupInfo, 
backupManager, e, "Unexpected Exception : ",
-173  BackupType.INCREMENTAL, 
conf);
-174throw new IOException(e);
-175  }
-176
-177}
-178  }
-179
-180  static class 
FullTableBackupClientForTest extends FullTableBackupClient
-181  {
-182
+139convertWALsToHFiles();
+140incrementalCopyHFiles(new 
String[] {getBulkOutputDir().toString()},
+141  
backupInfo.getBackupRootDir());
+142failStageIf(Stage.stage_2);
+143// Save list of WAL files 
copied
+144
backupManager.recordWALFiles(backupInfo.getIncrBackupFileList());
+145
+146// case INCR_BACKUP_COMPLETE:
+147// set overall backup status: 
complete. Here we make sure to complete the backup.
+148// After this checkpoint, even if 
entering cancel process, will let the backup finished
+149// Set the previousTimestampMap 
which is before this current log roll to the manifest.
+150HashMap> previousTimestampMap =
+151
backupManager.readLogTimestampMap();
+152
backupInfo.setIncrTimestampMap(previousTimestampMap);
+153
+154// The table list in backupInfo 
is good for both full backup and incremental backup.
+155// For incremental backup, it 
contains the incremental backup table set.
+156
backupManager.writeRegionServerLogTimestamp(backupInfo.getTables(), 
newTimestamps);
+157failStageIf(Stage.stage_3);
+158
+159HashMap> newTableSetTimestampMap =
+160
backupManager.readLogTimestampMap();
+161
+162Long newStartCode =
+163
BackupUtils.getMinValue(BackupUtils.getRSLogTimestampMins(newTableSetTimestampMap));
+164
backupManager.writeBackupStartCode(newStartCode);
+165
+166
handleBulkLoad(backupInfo.getTableNames());
+167failStageIf(Stage.stage_4);
+168
+169// backup complete
+170completeBackup(conn, backupInfo, 
backupManager, BackupType.INCREMENTAL, conf);
+171
+172  } catch (Exception e) {
+173failBackup(conn, backupInfo, 
backupManager, e, "Unexpected Exception : ",
+174  BackupType.INCREMENTAL, 
conf);
+175throw new IOException(e);
+176  }
+177
+178}
+179  }
+180
+181  static class 
FullTableBackupCl

[08/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-shaded-check-invariants/dependency-convergence.html
--
diff --git a/hbase-shaded-check-invariants/dependency-convergence.html 
b/hbase-shaded-check-invariants/dependency-convergence.html
index 9df25ff..1872ab1 100644
--- a/hbase-shaded-check-invariants/dependency-convergence.html
+++ b/hbase-shaded-check-invariants/dependency-convergence.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase Shaded Packaging Invariants – Reactor 
Dependency Convergence
 
@@ -123,10 +123,10 @@
 305
 
 Number of unique artifacts (NOA):
-331
+329
 
 Number of version-conflicting artifacts (NOC):
-17
+16
 
 Number of SNAPSHOT artifacts (NOS):
 0
@@ -191,75 +191,20 @@
 11.0.2
 
 
-org.apache.hbase:hbase-assembly:pom:3.0.0-SNAPSHOT+- org.apache.hbase:hbase-server:jar:3.0.0-SNAPSHOT:compile|  \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)\- org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-common:jar:2.7.1:compile|  \- (com.google.guava:guava:jar:11.0.2:compile - omitted for 
duplicate)\- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-client-project:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-client:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-endpoint:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-examples:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-mapreduce-client-core:jar:2.7.1:compile|  \- org.apache.hadoop:hadoop-yarn-common:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-hadoop-compat:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-metrics-api:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile  \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-hadoop2-compat:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-it:jar:3.0.0-SNAPSHOT+- org.apache.hbase:hbase-mapreduce:jar:3.0.0-SNAPSHOT:compile|  \- org.apache.hadoop:hadoop-hdfs:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)+- org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT:compile|  \- com.google.guava:guava:jar:11.0.2:compile\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- (com.google.guava:guava:jar:11.0.2:compile - 
omitted for duplicate)
-org.apache.hbase:hbase-mapreduce:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-common:jar:2.7.1:compile|  \- com.google.guava:guava:jar:11.0.2:compile+- org.apache.hadoop:hadoop-hdfs:jar:2.7.1:compile|  \- (com.google.guava:guava:jar:11.0.2:compile - omitted for 
duplicate)\- org.apache.hadoop:hadoop-minicluster:jar:2.7.1:test   +- org.apache.hadoop:hadoop-common:test-jar:tests:2.7.1:test   |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)   \- org.apache.hadoop:hadoop-yarn-server-tests:test-jar:tests:2.7.1:test  +- org.apache.hadoop:hadoop-yarn-server-nodemanager:jar:2.7.1:test  |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  +- org
 .apache.hadoop:hadoop-yarn-server-resourcemanager:jar:2.7.1:test  |  +- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  |  +- org.apache.hadoop:hadoop-yarn-server-applicationhistoryservice:jar:2.7.1:test  |  |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  |  \- org.apache.hadoop:hadoop-yarn-server-web-proxy:jar:2.7.1:test  | \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)
-org.apache.hbase:hbase-metrics-api:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-metrics:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:

[20/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-build-configuration/dependency-convergence.html
--
diff --git a/hbase-build-configuration/dependency-convergence.html 
b/hbase-build-configuration/dependency-convergence.html
index 688c4dc..bcda067 100644
--- a/hbase-build-configuration/dependency-convergence.html
+++ b/hbase-build-configuration/dependency-convergence.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Build Configuration – Reactor Dependency 
Convergence
 
@@ -123,10 +123,10 @@
 305
 
 Number of unique artifacts (NOA):
-331
+329
 
 Number of version-conflicting artifacts (NOC):
-17
+16
 
 Number of SNAPSHOT artifacts (NOS):
 0
@@ -191,75 +191,20 @@
 11.0.2
 
 
-org.apache.hbase:hbase-assembly:pom:3.0.0-SNAPSHOT+- org.apache.hbase:hbase-server:jar:3.0.0-SNAPSHOT:compile|  \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)\- org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-common:jar:2.7.1:compile|  \- (com.google.guava:guava:jar:11.0.2:compile - omitted for 
duplicate)\- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-client-project:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-client:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-endpoint:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-examples:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-mapreduce-client-core:jar:2.7.1:compile|  \- org.apache.hadoop:hadoop-yarn-common:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-hadoop-compat:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-metrics-api:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile  \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-hadoop2-compat:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-it:jar:3.0.0-SNAPSHOT+- org.apache.hbase:hbase-mapreduce:jar:3.0.0-SNAPSHOT:compile|  \- org.apache.hadoop:hadoop-hdfs:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)+- org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT:compile|  \- com.google.guava:guava:jar:11.0.2:compile\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- (com.google.guava:guava:jar:11.0.2:compile - 
omitted for duplicate)
-org.apache.hbase:hbase-mapreduce:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-common:jar:2.7.1:compile|  \- com.google.guava:guava:jar:11.0.2:compile+- org.apache.hadoop:hadoop-hdfs:jar:2.7.1:compile|  \- (com.google.guava:guava:jar:11.0.2:compile - omitted for 
duplicate)\- org.apache.hadoop:hadoop-minicluster:jar:2.7.1:test   +- org.apache.hadoop:hadoop-common:test-jar:tests:2.7.1:test   |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)   \- org.apache.hadoop:hadoop-yarn-server-tests:test-jar:tests:2.7.1:test  +- org.apache.hadoop:hadoop-yarn-server-nodemanager:jar:2.7.1:test  |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  +- org
 .apache.hadoop:hadoop-yarn-server-resourcemanager:jar:2.7.1:test  |  +- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  |  +- org.apache.hadoop:hadoop-yarn-server-applicationhistoryservice:jar:2.7.1:test  |  |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  |  \- org.apache.hadoop:hadoop-yarn-server-web-proxy:jar:2.7.1:test  | \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)
-org.apache.hbase:hbase-metrics-api:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-metrics:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.ap

[50/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/apidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.Operator.html
--
diff --git 
a/apidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.Operator.html 
b/apidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.Operator.html
index e48f3fd..fcfe8ab 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.Operator.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.Operator.html
@@ -29,536 +29,554 @@
 021import java.io.IOException;
 022import java.util.ArrayList;
 023import java.util.Arrays;
-024import java.util.Collections;
-025import java.util.List;
-026
-027import org.apache.hadoop.hbase.Cell;
-028import 
org.apache.hadoop.hbase.CellComparator;
-029import 
org.apache.hadoop.hbase.CellUtil;
-030import 
org.apache.yetus.audience.InterfaceAudience;
-031import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-032import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-033import 
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
-034import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
-035
-036/**
-037 * Implementation of {@link Filter} that 
represents an ordered List of Filters
-038 * which will be evaluated with a 
specified boolean operator {@link Operator#MUST_PASS_ALL}
-039 * (AND) or 
{@link Operator#MUST_PASS_ONE} (OR).
-040 * Since you can use Filter Lists as 
children of Filter Lists, you can create a
-041 * hierarchy of filters to be 
evaluated.
-042 *
-043 * 
-044 * {@link Operator#MUST_PASS_ALL} evaluates lazily: evaluation stops as soon as one filter does -045 * not include the KeyValue. -046 * -047 *
-048 * {@link Operator#MUST_PASS_ONE} evaluates non-lazily: all filters are always evaluated. -049 * -050 *
-051 * Defaults to {@link Operator#MUST_PASS_ALL}. -052 */ -053@InterfaceAudience.Public -054final public class FilterList extends FilterBase { -055 /** set operator */ -056 @InterfaceAudience.Public -057 public static enum Operator { -058/** !AND */ -059MUST_PASS_ALL, -060/** !OR */ -061MUST_PASS_ONE -062 } -063 -064 private static final int MAX_LOG_FILTERS = 5; -065 private Operator operator = Operator.MUST_PASS_ALL; -066 private final List filters; -067 private Filter seekHintFilter = null; -068 -069 /** Reference Cell used by {@link #transformCell(Cell)} for validation purpose. */ -070 private Cell referenceCell = null; -071 -072 /** -073 * When filtering a given Cell in {@link #filterKeyValue(Cell)}, -074 * this stores the transformed Cell to be returned by {@link #transformCell(Cell)}. -075 * -076 * Individual filters transformation are applied only when the filter includes the Cell. -077 * Transformations are composed in the order specified by {@link #filters}. -078 */ -079 private Cell transformedCell = null; -080 -081 /** -082 * Constructor that takes a set of {@link Filter}s. The default operator -083 * MUST_PASS_ALL is assumed. -084 * All filters are cloned to internal list. -085 * @param rowFilters list of filters -086 */ -087 public FilterList(final List rowFilters) { -088reversed = getReversed(rowFilters, reversed); -089this.filters = new ArrayList<>(rowFilters); -090 } -091 -092 /** -093 * Constructor that takes a var arg number of {@link Filter}s. The fefault operator -094 * MUST_PASS_ALL is assumed. -095 * @param rowFilters -096 */ -097 public FilterList(final Filter... rowFilters) { -098this(Arrays.asList(rowFilters)); -099 } -100 -101 /** -102 * Constructor that takes an operator. -103 * -104 * @param operator Operator to process filter set with. -105 */ -106 public FilterList(final Operator operator) { -107this.operator = operator; -108this.filters = new ArrayList<>(); -109 } -110 -111 /** -112 * Constructor that takes a set of {@link Filter}s and an operator. -113 * -114 * @param operator Operator to process filter set with. -115 * @param rowFilters Set of row filters. -116 */ -117 public FilterList(final Operator operator, final List rowFilters) { -118this(rowFilters); -119this.operator = operator; -120 } -121 -122 /** -123 * Constructor that takes a var arg number of {@link Filter}s and an operator. -124 * -125 * @param operator Operator to process filter set with. -126 * @param rowFilters Filters to use -127 */ -128 public FilterList(final Operator operator, final Filter... rowFilters) { -129this(rowFilters); -130this.operator = operator; -131 } -132 -133 /** -134 * Get the operator. -135 * -136 * @return operator -137 */ -138 public Operator getOperator() { -139return operator; -140 } -141 -142 /** -143 * Get the filters. -144 * -145 * @return filters -146 */ -147 public L

[09/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-build-configuration/hbase-spark/dependency-info.html
--
diff --git a/hbase-build-configuration/hbase-spark/dependency-info.html 
b/hbase-build-configuration/hbase-spark/dependency-info.html
index 92f2591..abd1422 100644
--- a/hbase-build-configuration/hbase-spark/dependency-info.html
+++ b/hbase-build-configuration/hbase-spark/dependency-info.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Spark – Dependency Information
 
@@ -147,7 +147,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-10-13
+  Last Published: 
2017-10-14
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-build-configuration/hbase-spark/dependency-management.html
--
diff --git a/hbase-build-configuration/hbase-spark/dependency-management.html 
b/hbase-build-configuration/hbase-spark/dependency-management.html
index 1962229..9d3917a 100644
--- a/hbase-build-configuration/hbase-spark/dependency-management.html
+++ b/hbase-build-configuration/hbase-spark/dependency-management.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Spark – Project Dependency Management
 
@@ -243,384 +243,390 @@
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>The Apache Software 
License, Version 2.0
 
+org.apache.curator
+http://curator.apache.org/curator-recipes";>curator-recipes
+4.0.0
+jar
+http://www.apache.org/licenses/LICENSE-2.0.txt";>The Apache Software 
License, Version 2.0
+
 org.apache.hadoop
 hadoop-auth
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-client
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-common
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-hdfs
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-mapreduce-client-core
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-mapreduce-client-jobclient
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-minicluster
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-annotations";>hbase-annotations
 3.0.0-SNAPSHOT
 test-jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-client";>hbase-client
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-common";>hbase-common
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-endpoint";>hbase-endpoint
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-support/hbase-error-prone";>hbase-error-prone
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-examples";>hbase-examples
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-external-blockcache";>hbase-external-blockcache
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-hadoop-compat";>hbase-hadoop-compat
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-hadoop2-compat";>hbase-hadoop2-compat
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-mapreduce";>hbase-mapreduce
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-metrics";>hbase-metrics
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/l

[45/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/dependency-convergence.html
--
diff --git a/dependency-convergence.html b/dependency-convergence.html
index 173e967..660e8a7 100644
--- a/dependency-convergence.html
+++ b/dependency-convergence.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase – Reactor Dependency Convergence
 
@@ -293,10 +293,10 @@
 305
 
 Number of unique artifacts (NOA):
-331
+329
 
 Number of version-conflicting artifacts (NOC):
-17
+16
 
 Number of SNAPSHOT artifacts (NOS):
 0
@@ -361,75 +361,20 @@
 11.0.2
 
 
-org.apache.hbase:hbase-assembly:pom:3.0.0-SNAPSHOT+- org.apache.hbase:hbase-server:jar:3.0.0-SNAPSHOT:compile|  \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)\- org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-common:jar:2.7.1:compile|  \- (com.google.guava:guava:jar:11.0.2:compile - omitted for 
duplicate)\- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-client-project:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-client:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-endpoint:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-examples:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-mapreduce-client-core:jar:2.7.1:compile|  \- org.apache.hadoop:hadoop-yarn-common:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-hadoop-compat:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-metrics-api:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile  \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-hadoop2-compat:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-it:jar:3.0.0-SNAPSHOT+- org.apache.hbase:hbase-mapreduce:jar:3.0.0-SNAPSHOT:compile|  \- org.apache.hadoop:hadoop-hdfs:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)+- org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT:compile|  \- com.google.guava:guava:jar:11.0.2:compile\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- (com.google.guava:guava:jar:11.0.2:compile - 
omitted for duplicate)
-org.apache.hbase:hbase-mapreduce:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-common:jar:2.7.1:compile|  \- com.google.guava:guava:jar:11.0.2:compile+- org.apache.hadoop:hadoop-hdfs:jar:2.7.1:compile|  \- (com.google.guava:guava:jar:11.0.2:compile - omitted for 
duplicate)\- org.apache.hadoop:hadoop-minicluster:jar:2.7.1:test   +- org.apache.hadoop:hadoop-common:test-jar:tests:2.7.1:test   |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)   \- org.apache.hadoop:hadoop-yarn-server-tests:test-jar:tests:2.7.1:test  +- org.apache.hadoop:hadoop-yarn-server-nodemanager:jar:2.7.1:test  |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  +- org
 .apache.hadoop:hadoop-yarn-server-resourcemanager:jar:2.7.1:test  |  +- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  |  +- org.apache.hadoop:hadoop-yarn-server-applicationhistoryservice:jar:2.7.1:test  |  |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  |  \- org.apache.hadoop:hadoop-yarn-server-web-proxy:jar:2.7.1:test  | \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)
-org.apache.hbase:hbase-metrics-api:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-metrics:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-procedure:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apac

[07/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-shaded-check-invariants/dependency-info.html
--
diff --git a/hbase-shaded-check-invariants/dependency-info.html 
b/hbase-shaded-check-invariants/dependency-info.html
index 345a80e..c06ad6d 100644
--- a/hbase-shaded-check-invariants/dependency-info.html
+++ b/hbase-shaded-check-invariants/dependency-info.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase Shaded Packaging Invariants – Dependency 
Information
 
@@ -148,7 +148,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-10-13
+  Last Published: 
2017-10-14
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-shaded-check-invariants/dependency-management.html
--
diff --git a/hbase-shaded-check-invariants/dependency-management.html 
b/hbase-shaded-check-invariants/dependency-management.html
index 69a6b47..ecfa1ae 100644
--- a/hbase-shaded-check-invariants/dependency-management.html
+++ b/hbase-shaded-check-invariants/dependency-management.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase Shaded Packaging Invariants – Project 
Dependency Management
 
@@ -243,384 +243,390 @@
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>The Apache Software 
License, Version 2.0
 
+org.apache.curator
+http://curator.apache.org/curator-recipes";>curator-recipes
+4.0.0
+jar
+http://www.apache.org/licenses/LICENSE-2.0.txt";>The Apache Software 
License, Version 2.0
+
 org.apache.hadoop
 hadoop-auth
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-client
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-common
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-hdfs
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-mapreduce-client-core
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-mapreduce-client-jobclient
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-minicluster
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-annotations";>hbase-annotations
 3.0.0-SNAPSHOT
 test-jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-client";>hbase-client
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-common";>hbase-common
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-endpoint";>hbase-endpoint
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-support/hbase-error-prone";>hbase-error-prone
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-examples";>hbase-examples
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-external-blockcache";>hbase-external-blockcache
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-hadoop-compat";>hbase-hadoop-compat
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-hadoop2-compat";>hbase-hadoop2-compat
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-mapreduce";>hbase-mapreduce
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-metrics";>hbase-metrics
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache Licen

[10/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-build-configuration/hbase-spark/dependency-convergence.html
--
diff --git a/hbase-build-configuration/hbase-spark/dependency-convergence.html 
b/hbase-build-configuration/hbase-spark/dependency-convergence.html
index 844913b..1d6fee5 100644
--- a/hbase-build-configuration/hbase-spark/dependency-convergence.html
+++ b/hbase-build-configuration/hbase-spark/dependency-convergence.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Spark – Reactor Dependency Convergence
 
@@ -123,10 +123,10 @@
 305
 
 Number of unique artifacts (NOA):
-331
+329
 
 Number of version-conflicting artifacts (NOC):
-17
+16
 
 Number of SNAPSHOT artifacts (NOS):
 0
@@ -191,75 +191,20 @@
 11.0.2
 
 
-org.apache.hbase:hbase-assembly:pom:3.0.0-SNAPSHOT+- org.apache.hbase:hbase-server:jar:3.0.0-SNAPSHOT:compile|  \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)\- org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-common:jar:2.7.1:compile|  \- (com.google.guava:guava:jar:11.0.2:compile - omitted for 
duplicate)\- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-client-project:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-client:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-endpoint:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-examples:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-mapreduce-client-core:jar:2.7.1:compile|  \- org.apache.hadoop:hadoop-yarn-common:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-hadoop-compat:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-metrics-api:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile  \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-hadoop2-compat:jar:3.0.0-SNAPSHOT\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-it:jar:3.0.0-SNAPSHOT+- org.apache.hbase:hbase-mapreduce:jar:3.0.0-SNAPSHOT:compile|  \- org.apache.hadoop:hadoop-hdfs:jar:2.7.1:compile| \- (com.google.guava:guava:jar:11.0.2:compile
 - omitted for duplicate)+- org.apache.hbase:hbase-backup:jar:3.0.0-SNAPSHOT:compile|  \- com.google.guava:guava:jar:11.0.2:compile\- org.apache.hadoop:hadoop-common:jar:2.7.1:compile   \- (com.google.guava:guava:jar:11.0.2:compile - 
omitted for duplicate)
-org.apache.hbase:hbase-mapreduce:jar:3.0.0-SNAPSHOT+- org.apache.hadoop:hadoop-common:jar:2.7.1:compile|  \- com.google.guava:guava:jar:11.0.2:compile+- org.apache.hadoop:hadoop-hdfs:jar:2.7.1:compile|  \- (com.google.guava:guava:jar:11.0.2:compile - omitted for 
duplicate)\- org.apache.hadoop:hadoop-minicluster:jar:2.7.1:test   +- org.apache.hadoop:hadoop-common:test-jar:tests:2.7.1:test   |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)   \- org.apache.hadoop:hadoop-yarn-server-tests:test-jar:tests:2.7.1:test  +- org.apache.hadoop:hadoop-yarn-server-nodemanager:jar:2.7.1:test  |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  +- org
 .apache.hadoop:hadoop-yarn-server-resourcemanager:jar:2.7.1:test  |  +- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  |  +- org.apache.hadoop:hadoop-yarn-server-applicationhistoryservice:jar:2.7.1:test  |  |  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  |  \- org.apache.hadoop:hadoop-yarn-server-web-proxy:jar:2.7.1:test  | \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)  \- (com.google.guava:guava:jar:11.0.2:test
 - omitted for duplicate)
-org.apache.hbase:hbase-metrics-api:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.google.guava:guava:jar:11.0.2:compile
-org.apache.hbase:hbase-metrics:jar:3.0.0-SNAPSHOT\- org.apache.hbase:hbase-common:jar:3.0.0-SNAPSHOT:compile   \- org.apache.hadoop:hadoop-common:jar:2.7.1:compile  \- com.

[24/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/StoreScanner.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/StoreScanner.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/StoreScanner.html
index b2a8fdc..de07779 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/StoreScanner.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/StoreScanner.html
@@ -511,647 +511,642 @@
 503return this.heap.seek(key);
 504  }
 505
-506  @Override
-507  public boolean next(List 
outResult) throws IOException {
-508return next(outResult, 
NoLimitScannerContext.getInstance());
-509  }
-510
-511  /**
-512   * Get the next row of values from this 
Store.
-513   * @param outResult
-514   * @param scannerContext
-515   * @return true if there are more rows, 
false if scanner is done
-516   */
-517  @Override
-518  public boolean next(List 
outResult, ScannerContext scannerContext) throws IOException {
-519if (scannerContext == null) {
-520  throw new 
IllegalArgumentException("Scanner context cannot be null");
-521}
-522if (checkFlushed() && 
reopenAfterFlush()) {
-523  return 
scannerContext.setScannerState(NextState.MORE_VALUES).hasMoreValues();
-524}
-525
-526// if the heap was left null, then 
the scanners had previously run out anyways, close and
-527// return.
-528if (this.heap == null) {
-529  // By this time partial close 
should happened because already heap is null
-530  close(false);// Do all cleanup 
except heap.close()
-531  return 
scannerContext.setScannerState(NextState.NO_MORE_VALUES).hasMoreValues();
-532}
-533
-534Cell cell = this.heap.peek();
-535if (cell == null) {
-536  close(false);// Do all cleanup 
except heap.close()
-537  return 
scannerContext.setScannerState(NextState.NO_MORE_VALUES).hasMoreValues();
-538}
-539
-540// only call setRow if the row 
changes; avoids confusing the query matcher
-541// if scanning intra-row
-542
-543// If no limits exists in the scope 
LimitScope.Between_Cells then we are sure we are changing
-544// rows. Else it is possible we are 
still traversing the same row so we must perform the row
-545// comparison.
-546if 
(!scannerContext.hasAnyLimit(LimitScope.BETWEEN_CELLS) || matcher.currentRow() 
== null) {
-547  this.countPerRow = 0;
-548  matcher.setToNewRow(cell);
+506  /**
+507   * Get the next row of values from this 
Store.
+508   * @param outResult
+509   * @param scannerContext
+510   * @return true if there are more rows, 
false if scanner is done
+511   */
+512  @Override
+513  public boolean next(List 
outResult, ScannerContext scannerContext) throws IOException {
+514if (scannerContext == null) {
+515  throw new 
IllegalArgumentException("Scanner context cannot be null");
+516}
+517if (checkFlushed() && 
reopenAfterFlush()) {
+518  return 
scannerContext.setScannerState(NextState.MORE_VALUES).hasMoreValues();
+519}
+520
+521// if the heap was left null, then 
the scanners had previously run out anyways, close and
+522// return.
+523if (this.heap == null) {
+524  // By this time partial close 
should happened because already heap is null
+525  close(false);// Do all cleanup 
except heap.close()
+526  return 
scannerContext.setScannerState(NextState.NO_MORE_VALUES).hasMoreValues();
+527}
+528
+529Cell cell = this.heap.peek();
+530if (cell == null) {
+531  close(false);// Do all cleanup 
except heap.close()
+532  return 
scannerContext.setScannerState(NextState.NO_MORE_VALUES).hasMoreValues();
+533}
+534
+535// only call setRow if the row 
changes; avoids confusing the query matcher
+536// if scanning intra-row
+537
+538// If no limits exists in the scope 
LimitScope.Between_Cells then we are sure we are changing
+539// rows. Else it is possible we are 
still traversing the same row so we must perform the row
+540// comparison.
+541if 
(!scannerContext.hasAnyLimit(LimitScope.BETWEEN_CELLS) || matcher.currentRow() 
== null) {
+542  this.countPerRow = 0;
+543  matcher.setToNewRow(cell);
+544}
+545
+546// Clear progress away unless invoker 
has indicated it should be kept.
+547if 
(!scannerContext.getKeepProgress()) {
+548  scannerContext.clearProgress();
 549}
 550
-551// Clear progress away unless invoker 
has indicated it should be kept.
-552if 
(!scannerContext.getKeepProgress()) {
-553  scannerContext.clearProgress();
-554}
-555
-556// Only do a sanity-check if store 
and comparator are available.
-557CellComparator comparator = 
store.map(s -> s.getComparator()).orElse(null);
-558
-559int count = 0;
-560long totalBytesRead = 0;
-561
-562LOOP: do {
-563  // Upda

[21/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-annotations/dependency-info.html
--
diff --git a/hbase-annotations/dependency-info.html 
b/hbase-annotations/dependency-info.html
index 073ef76..164b100 100644
--- a/hbase-annotations/dependency-info.html
+++ b/hbase-annotations/dependency-info.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Annotations – Dependency Information
 
@@ -147,7 +147,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-10-13
+  Last Published: 
2017-10-14
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-annotations/dependency-management.html
--
diff --git a/hbase-annotations/dependency-management.html 
b/hbase-annotations/dependency-management.html
index ea7532f..54fd1c0 100644
--- a/hbase-annotations/dependency-management.html
+++ b/hbase-annotations/dependency-management.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Annotations – Project Dependency 
Management
 
@@ -243,384 +243,390 @@
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>The Apache Software 
License, Version 2.0
 
+org.apache.curator
+http://curator.apache.org/curator-recipes";>curator-recipes
+4.0.0
+jar
+http://www.apache.org/licenses/LICENSE-2.0.txt";>The Apache Software 
License, Version 2.0
+
 org.apache.hadoop
 hadoop-auth
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-client
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-common
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-hdfs
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-mapreduce-client-core
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-mapreduce-client-jobclient
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-minicluster
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-annotations";>hbase-annotations
 3.0.0-SNAPSHOT
 test-jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-client";>hbase-client
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-common";>hbase-common
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-endpoint";>hbase-endpoint
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-support/hbase-error-prone";>hbase-error-prone
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-examples";>hbase-examples
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-external-blockcache";>hbase-external-blockcache
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-hadoop-compat";>hbase-hadoop-compat
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-hadoop2-compat";>hbase-hadoop2-compat
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-mapreduce";>hbase-mapreduce
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-metrics";>hbase-metrics
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-metrics-api";>hbase-metrics-api
 3.0.0-SNAPSHOT
 jar

[19/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-build-configuration/dependency-info.html
--
diff --git a/hbase-build-configuration/dependency-info.html 
b/hbase-build-configuration/dependency-info.html
index 12d469c..930e823 100644
--- a/hbase-build-configuration/dependency-info.html
+++ b/hbase-build-configuration/dependency-info.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Build Configuration – Dependency 
Information
 
@@ -148,7 +148,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-10-13
+  Last Published: 
2017-10-14
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-build-configuration/dependency-management.html
--
diff --git a/hbase-build-configuration/dependency-management.html 
b/hbase-build-configuration/dependency-management.html
index 9937134..7fc7fb1 100644
--- a/hbase-build-configuration/dependency-management.html
+++ b/hbase-build-configuration/dependency-management.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Build Configuration – Project Dependency 
Management
 
@@ -243,384 +243,390 @@
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>The Apache Software 
License, Version 2.0
 
+org.apache.curator
+http://curator.apache.org/curator-recipes";>curator-recipes
+4.0.0
+jar
+http://www.apache.org/licenses/LICENSE-2.0.txt";>The Apache Software 
License, Version 2.0
+
 org.apache.hadoop
 hadoop-auth
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-client
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-common
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-hdfs
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-mapreduce-client-core
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-mapreduce-client-jobclient
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-minicluster
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-annotations";>hbase-annotations
 3.0.0-SNAPSHOT
 test-jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-client";>hbase-client
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-common";>hbase-common
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-endpoint";>hbase-endpoint
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-support/hbase-error-prone";>hbase-error-prone
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-examples";>hbase-examples
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-external-blockcache";>hbase-external-blockcache
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-hadoop-compat";>hbase-hadoop-compat
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-hadoop2-compat";>hbase-hadoop2-compat
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-mapreduce";>hbase-mapreduce
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-metrics";>hbase-metrics
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase

[15/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-build-configuration/hbase-archetypes/hbase-archetype-builder/dependency-info.html
--
diff --git 
a/hbase-build-configuration/hbase-archetypes/hbase-archetype-builder/dependency-info.html
 
b/hbase-build-configuration/hbase-archetypes/hbase-archetype-builder/dependency-info.html
index b850c0a..faa2318 100644
--- 
a/hbase-build-configuration/hbase-archetypes/hbase-archetype-builder/dependency-info.html
+++ 
b/hbase-build-configuration/hbase-archetypes/hbase-archetype-builder/dependency-info.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Archetype builder – Dependency 
Information
 
@@ -148,7 +148,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-10-13
+  Last Published: 
2017-10-14
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-build-configuration/hbase-archetypes/hbase-archetype-builder/dependency-management.html
--
diff --git 
a/hbase-build-configuration/hbase-archetypes/hbase-archetype-builder/dependency-management.html
 
b/hbase-build-configuration/hbase-archetypes/hbase-archetype-builder/dependency-management.html
index 901826c..b02d628 100644
--- 
a/hbase-build-configuration/hbase-archetypes/hbase-archetype-builder/dependency-management.html
+++ 
b/hbase-build-configuration/hbase-archetypes/hbase-archetype-builder/dependency-management.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Archetype builder – Project Dependency 
Management
 
@@ -243,384 +243,390 @@
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>The Apache Software 
License, Version 2.0
 
+org.apache.curator
+http://curator.apache.org/curator-recipes";>curator-recipes
+4.0.0
+jar
+http://www.apache.org/licenses/LICENSE-2.0.txt";>The Apache Software 
License, Version 2.0
+
 org.apache.hadoop
 hadoop-auth
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-client
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-common
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-hdfs
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-mapreduce-client-core
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-mapreduce-client-jobclient
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-minicluster
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-annotations";>hbase-annotations
 3.0.0-SNAPSHOT
 test-jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-client";>hbase-client
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-common";>hbase-common
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-endpoint";>hbase-endpoint
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-support/hbase-error-prone";>hbase-error-prone
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-examples";>hbase-examples
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-external-blockcache";>hbase-external-blockcache
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-hadoop-compat";>hbase-hadoop-compat
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-hadoop2-compat";>hbase-hadoop2-compat
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hba

[51/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
Published site at .


Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/7816cbde
Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/7816cbde
Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/7816cbde

Branch: refs/heads/asf-site
Commit: 7816cbdea8a38d01aa76f1a1fe43213f50356cce
Parents: c1db450
Author: jenkins 
Authored: Sat Oct 14 15:15:24 2017 +
Committer: jenkins 
Committed: Sat Oct 14 15:15:24 2017 +

--
 acid-semantics.html | 4 +-
 apache_hbase_reference_guide.pdf| 4 +-
 .../hbase/filter/FilterList.Operator.html   |10 +-
 .../apache/hadoop/hbase/filter/FilterList.html  |54 +-
 .../hbase/filter/FilterList.Operator.html   |  1062 +-
 .../apache/hadoop/hbase/filter/FilterList.html  |  1062 +-
 book.html   | 2 +-
 bulk-loads.html | 4 +-
 checkstyle-aggregate.html   | 22632 +
 checkstyle.rss  |26 +-
 coc.html| 4 +-
 cygwin.html | 4 +-
 dependencies.html   | 4 +-
 dependency-convergence.html |   156 +-
 dependency-info.html| 4 +-
 dependency-management.html  |   136 +-
 devapidocs/allclasses-frame.html| 3 +-
 devapidocs/allclasses-noframe.html  | 3 +-
 devapidocs/constant-values.html |35 +-
 devapidocs/index-all.html   |   104 +-
 .../hbase/backup/class-use/BackupInfo.html  |18 +-
 .../impl/IncrementalTableBackupClient.html  |88 +-
 .../MapReduceBackupCopyJob.BackupDistCp.html|   107 +-
 .../MapReduceBackupCopyJob.SnapshotCopy.html|12 +-
 .../mapreduce/MapReduceBackupCopyJob.html   |49 +-
 .../hadoop/hbase/backup/package-tree.html   | 4 +-
 .../org/apache/hadoop/hbase/class-use/Cell.html |25 +-
 .../hadoop/hbase/class-use/Coprocessor.html | 4 +-
 .../hbase/class-use/CoprocessorEnvironment.html | 2 +-
 .../hadoop/hbase/client/class-use/Scan.html |51 +-
 .../hadoop/hbase/client/package-tree.html   |26 +-
 .../apache/hadoop/hbase/client/package-use.html | 5 -
 .../coprocessor/class-use/ObserverContext.html  |32 +-
 .../class-use/RegionCoprocessor.html| 4 +-
 .../class-use/RegionCoprocessorEnvironment.html |46 +-
 .../coprocessor/class-use/RegionObserver.html   | 4 +-
 .../example/BulkDeleteEndpoint.Column.html  | 4 +-
 .../example/DelegatingInternalScanner.html  |   366 +
 .../ExampleMasterObserverWithMetrics.html   | 4 +-
 ...ooKeeperScanPolicyObserver.ZKDataHolder.html |   416 +
 .../ZooKeeperScanPolicyObserver.ZKWatcher.html  |   402 -
 .../example/ZooKeeperScanPolicyObserver.html|   300 +-
 .../class-use/DelegatingInternalScanner.html|   125 +
 ...ooKeeperScanPolicyObserver.ZKDataHolder.html |   125 +
 .../ZooKeeperScanPolicyObserver.ZKWatcher.html  |   125 -
 .../coprocessor/example/package-frame.html  | 3 +-
 .../coprocessor/example/package-summary.html|22 +-
 .../hbase/coprocessor/example/package-tree.html | 3 +-
 .../hadoop/hbase/executor/package-tree.html | 2 +-
 .../hbase/filter/FilterList.Operator.html   |10 +-
 .../apache/hadoop/hbase/filter/FilterList.html  |82 +-
 .../hadoop/hbase/filter/class-use/Filter.html   | 8 +-
 .../hadoop/hbase/filter/package-tree.html   | 8 +-
 .../hadoop/hbase/io/hfile/package-tree.html | 6 +-
 .../apache/hadoop/hbase/ipc/package-tree.html   | 2 +-
 .../hadoop/hbase/mapreduce/package-tree.html| 4 +-
 .../hadoop/hbase/master/package-tree.html   | 4 +-
 .../hbase/master/procedure/package-tree.html| 2 +-
 .../org/apache/hadoop/hbase/package-tree.html   |12 +-
 .../hadoop/hbase/procedure2/package-tree.html   | 4 +-
 .../hadoop/hbase/quotas/package-tree.html   | 8 +-
 .../hbase/regionserver/InternalScanner.html |22 +-
 .../KeyValueHeap.KVScannerComparator.html   |12 +-
 .../hadoop/hbase/regionserver/KeyValueHeap.html |79 +-
 .../hbase/regionserver/MobStoreScanner.html | 9 +-
 .../regionserver/ReversedKeyValueHeap.html  | 9 +-
 .../regionserver/ReversedMobStoreScanner.html   | 9 +-
 .../regionserver/ReversedStoreScanner.html  | 9 +-
 .../hadoop/hbase/regionserver/StoreScanner.html |   123 +-
 .../regionserver/class-use/InternalScanner.html |93 +-
 .../regionserver/class-use/KeyValueScanner.html |78 +-
 .../hbase/regionserver/class-use/ScanInfo.html  |25 +-
 .../hbase/regionserver/class-use/ScanType.html  |15 +-
 .../regionserver/class-

[34/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/IncrementalTableBackupClient.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/IncrementalTableBackupClient.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/IncrementalTableBackupClient.html
index bcf966e..de32a91 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/IncrementalTableBackupClient.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/impl/IncrementalTableBackupClient.html
@@ -26,100 +26,100 @@
 018
 019package 
org.apache.hadoop.hbase.backup.impl;
 020
-021import java.io.FileNotFoundException;
-022import java.io.IOException;
-023import java.net.URI;
-024import java.net.URISyntaxException;
-025import java.util.ArrayList;
-026import java.util.HashMap;
-027import java.util.List;
-028import java.util.Map;
-029import java.util.Set;
-030import java.util.TreeMap;
-031
-032import 
org.apache.commons.lang3.StringUtils;
-033import org.apache.commons.logging.Log;
-034import 
org.apache.commons.logging.LogFactory;
-035import org.apache.hadoop.fs.FileSystem;
-036import org.apache.hadoop.fs.FileUtil;
-037import org.apache.hadoop.fs.Path;
-038import 
org.apache.hadoop.hbase.TableName;
-039import 
org.apache.hadoop.hbase.backup.BackupCopyJob;
-040import 
org.apache.hadoop.hbase.backup.BackupInfo;
-041import 
org.apache.hadoop.hbase.backup.BackupInfo.BackupPhase;
-042import 
org.apache.hadoop.hbase.backup.BackupRequest;
-043import 
org.apache.hadoop.hbase.backup.BackupRestoreFactory;
-044import 
org.apache.hadoop.hbase.backup.BackupType;
-045import 
org.apache.hadoop.hbase.backup.util.BackupUtils;
-046import 
org.apache.yetus.audience.InterfaceAudience;
-047import 
org.apache.hadoop.hbase.client.Admin;
-048import 
org.apache.hadoop.hbase.client.Connection;
-049import 
org.apache.hadoop.hbase.mapreduce.WALPlayer;
-050import 
org.apache.hadoop.hbase.util.Bytes;
-051import 
org.apache.hadoop.hbase.util.FSUtils;
-052import 
org.apache.hadoop.hbase.util.HFileArchiveUtil;
-053import 
org.apache.hadoop.hbase.util.Pair;
-054import 
org.apache.hadoop.hbase.wal.AbstractFSWALProvider;
-055import org.apache.hadoop.util.Tool;
-056
-057/**
-058 * Incremental backup implementation.
-059 * See the {@link #execute() execute} 
method.
-060 *
-061 */
-062@InterfaceAudience.Private
-063public class IncrementalTableBackupClient 
extends TableBackupClient {
-064  private static final Log LOG = 
LogFactory.getLog(IncrementalTableBackupClient.class);
-065
-066  protected 
IncrementalTableBackupClient() {
-067  }
-068
-069  public 
IncrementalTableBackupClient(final Connection conn, final String backupId,
-070  BackupRequest request) throws 
IOException {
-071super(conn, backupId, request);
-072  }
-073
-074  protected List 
filterMissingFiles(List incrBackupFileList) throws IOException 
{
-075FileSystem fs = 
FileSystem.get(conf);
-076List list = new 
ArrayList();
-077for (String file : 
incrBackupFileList) {
-078  Path p = new Path(file);
-079  if (fs.exists(p) || 
isActiveWalPath(p)) {
-080list.add(file);
-081  } else {
-082LOG.warn("Can't find file: " + 
file);
-083  }
-084}
-085return list;
-086  }
-087
-088  /**
-089   * Check if a given path is belongs to 
active WAL directory
-090   * @param p path
-091   * @return true, if yes
-092   */
-093  protected boolean isActiveWalPath(Path 
p) {
-094return 
!AbstractFSWALProvider.isArchivedLogFile(p);
-095  }
-096
-097  protected static int getIndex(TableName 
tbl, List sTableList) {
-098if (sTableList == null) return 0;
-099for (int i = 0; i < 
sTableList.size(); i++) {
-100  if (tbl.equals(sTableList.get(i))) 
{
-101return i;
-102  }
-103}
-104return -1;
-105  }
-106
-107  /*
-108   * Reads bulk load records from backup 
table, iterates through the records and forms the paths
-109   * for bulk loaded hfiles. Copies the 
bulk loaded hfiles to backup destination
-110   * @param sTableList list of tables to 
be backed up
-111   * @return map of table to List of 
files
-112   */
-113  protected Map>[] handleBulkLoad(List sTableList) throws 
IOException {
-114Map>[] 
mapForSrc = new Map[sTableList.size()];
+021import java.io.IOException;
+022import java.net.URI;
+023import java.net.URISyntaxException;
+024import java.util.ArrayList;
+025import java.util.HashMap;
+026import java.util.List;
+027import java.util.Map;
+028import java.util.Set;
+029import java.util.TreeMap;
+030
+031import 
org.apache.commons.lang3.StringUtils;
+032import org.apache.commons.logging.Log;
+033import 
org.apache.commons.logging.LogFactory;
+034import org.apache.hadoop.fs.FileSystem;
+035import org.apache.hadoop.fs.Path;
+036import 
org.apache.hadoop.hbase.TableName;
+0

[42/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/org/apache/hadoop/hbase/backup/impl/IncrementalTableBackupClient.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/backup/impl/IncrementalTableBackupClient.html
 
b/devapidocs/org/apache/hadoop/hbase/backup/impl/IncrementalTableBackupClient.html
index 10c3c4b..96d5c9e 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/backup/impl/IncrementalTableBackupClient.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/backup/impl/IncrementalTableBackupClient.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":9,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":9,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -115,7 +115,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Private
-public class IncrementalTableBackupClient
+public class IncrementalTableBackupClient
 extends TableBackupClient
 Incremental backup implementation.
  See the execute
 method.
@@ -205,55 +205,61 @@ extends 
 protected void
-convertWALsToHFiles(BackupInfo backupInfo) 
+convertWALsToHFiles() 
 
 
+private void
+copyBulkLoadedFiles(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">ListString> activeFiles,
+   http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">ListString> archiveFiles) 
+
+
 protected void
 deleteBulkLoadDirectory() 
 
-
+
 void
 execute()
 Backup request execution
 
 
-
+
 protected http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">ListString>
 filterMissingFiles(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">ListString> incrBackupFileList) 
 
-
+
 protected org.apache.hadoop.fs.Path
 getBulkOutputDir() 
 
-
+
 protected org.apache.hadoop.fs.Path
 getBulkOutputDirForTable(TableName table) 
 
-
+
 protected static int
 getIndex(TableName tbl,
 http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List sTableList) 
 
-
+
 protected http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true";
 title="class or interface in java.util">MapList>[]
 handleBulkLoad(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List sTableList) 
 
-
+
 protected void
-incrementalCopyHFiles(BackupInfo backupInfo) 
+incrementalCopyHFiles(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String[] files,
+ http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in 
java.lang">String backupDest) 
 
-
+
 protected boolean
 isActiveWalPath(org.apache.hadoop.fs.Path p)
 Check if a given path is belongs to active WAL 
directory
 
 
-
+
 protected boolean
 tableExists(TableName table,
Connection conn) 
 
-
+
 protected void
 walToHFiles(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">ListString> dirPaths,
TableName tableName) 
@@ -293,7 +299,7 @@ extends 
 
 LOG
-private static final org.apache.commons.logging.Log LOG
+private static final org.apache.commons.logging.Log LOG
 
 
 
@@ -310,7 +316,7 @@ extends 
 
 IncrementalTableBackupClient
-protected IncrementalTableBackupClient()
+protected IncrementalTableBackupClient()
 
 
 
@@ -319,7 +325,7 @@ extends 
 
 IncrementalTableBackupClient
-public IncrementalTableBackupClient(Connection conn,
+public IncrementalTableBackupClient(Connection conn,

[06/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/testdevapidocs/index-all.html
--
diff --git a/testdevapidocs/index-all.html b/testdevapidocs/index-all.html
index d05c6ef..5ab97a8 100644
--- a/testdevapidocs/index-all.html
+++ b/testdevapidocs/index-all.html
@@ -1583,6 +1583,8 @@
 
 Assert that we don't have any snapshots lists
 
+assertNotExists(int,
 int) - Method in class 
org.apache.hadoop.hbase.coprocessor.example.TestZooKeeperScanPolicyObserver
+ 
 assertNoThreadsMatching(String)
 - Static method in class org.apache.hadoop.hbase.GenericTestUtils
 
 Assert that there are no threads running whose name matches 
the
@@ -1775,6 +1777,8 @@
  
 assertUserGroup(User,
 ImmutableSet) - Method in class 
org.apache.hadoop.hbase.security.TestUser
  
+assertValueEquals(int,
 int) - Method in class 
org.apache.hadoop.hbase.coprocessor.example.TestZooKeeperScanPolicyObserver
+ 
 assertValueNear(long,
 long, long) - Static method in class org.apache.hadoop.hbase.GenericTestUtils
  
 assertValueSize(int,
 int) - Static method in class org.apache.hadoop.hbase.PerformanceEvaluationCommons
@@ -10789,6 +10793,8 @@
  
 FAMILY
 - Static variable in class org.apache.hadoop.hbase.coprocessor.example.TestRefreshHFilesEndpoint
  
+FAMILY
 - Static variable in class org.apache.hadoop.hbase.coprocessor.example.TestZooKeeperScanPolicyObserver
+ 
 family
 - Static variable in class org.apache.hadoop.hbase.coprocessor.TestNegativeMemStoreSizeWithSlowCoprocessor
  
 family
 - Static variable in class org.apache.hadoop.hbase.coprocessor.TestOpenTableInCoprocessor
@@ -22697,6 +22703,8 @@
  
 name
 - Variable in class org.apache.hadoop.hbase.constraint.TestConstraints
  
+NAME
 - Static variable in class org.apache.hadoop.hbase.coprocessor.example.TestZooKeeperScanPolicyObserver
+ 
 name
 - Variable in class org.apache.hadoop.hbase.coprocessor.TestCoprocessorInterface
  
 name
 - Variable in class org.apache.hadoop.hbase.coprocessor.TestCoprocessorMetrics
@@ -23337,12 +23345,8 @@
  
 next(long,
 int) - Method in class org.apache.hadoop.hbase.master.MockRegionServer
  
-next(List)
 - Method in class org.apache.hadoop.hbase.regionserver.compactions.TestCompactor.Scanner
- 
 next(List,
 ScannerContext) - Method in class 
org.apache.hadoop.hbase.regionserver.compactions.TestCompactor.Scanner
  
-next(List)
 - Method in class org.apache.hadoop.hbase.regionserver.compactions.TestStripeCompactionPolicy.Scanner
- 
 next(List,
 ScannerContext) - Method in class 
org.apache.hadoop.hbase.regionserver.compactions.TestStripeCompactionPolicy.Scanner
  
 next()
 - Method in class org.apache.hadoop.hbase.regionserver.DelegatingKeyValueScanner
@@ -27225,6 +27229,8 @@
  
 put - 
Variable in class org.apache.hadoop.hbase.client.TestPutDotHas
  
+put(int,
 int, long) - Method in class 
org.apache.hadoop.hbase.coprocessor.example.TestZooKeeperScanPolicyObserver
+ 
 PUT
 - Static variable in class org.apache.hadoop.hbase.quotas.policies.BaseViolationPolicyEnforcement
  
 put(byte[],
 Put) - Method in class org.apache.hadoop.hbase.regionserver.OOMERegionServer
@@ -27581,6 +27587,8 @@
  
 QUALIFIER
 - Static variable in class org.apache.hadoop.hbase.coprocessor.example.TestRefreshHFilesEndpoint
  
+QUALIFIER
 - Static variable in class org.apache.hadoop.hbase.coprocessor.example.TestZooKeeperScanPolicyObserver
+ 
 qualifier
 - Static variable in class org.apache.hadoop.hbase.coprocessor.TestNegativeMemStoreSizeWithSlowCoprocessor
  
 QUALIFIER
 - Static variable in class org.apache.hadoop.hbase.filter.TestDependentColumnFilter
@@ -31847,6 +31855,8 @@
  
 setExceptionInjector(TestMetaCache.ExceptionInjector)
 - Method in class org.apache.hadoop.hbase.client.TestMetaCache.RegionServerWithFakeRpcServices
  
+setExpireBefore(long)
 - Method in class org.apache.hadoop.hbase.coprocessor.example.TestZooKeeperScanPolicyObserver
+ 
 setFailures(TestBackupDeleteWithFailures.Failure...)
 - Method in class org.apache.hadoop.hbase.backup.TestBackupDeleteWithFailures.MasterSnapshotObserver
  
 setFailureType(FaultyFSLog.FailureType)
 - Method in class org.apache.hadoop.hbase.wal.FaultyFSLog
@@ -32284,6 +32294,8 @@
  
 setUp(String)
 - Static method in class org.apache.hadoop.hbase.coprocessor.example.TestRefreshHFilesEndpoint
  
+setUp()
 - Static method in class org.apache.hadoop.hbase.coprocessor.example.TestZooKeeperScanPolicyObserver
+ 
 setup()
 - Method in class org.apache.hadoop.hbase.coprocessor.TestCoprocessorMetrics
  
 setUp()
 - Method in class org.apache.hadoop.hbase.coprocessor.TestRegionObserverBypass
@@ -36186,6 +36198,8 @@
  
 table
 - Static variable in class org.apache.hadoop.hbase.coprocessor.example.TestRefreshHFilesEndpoint
  
+TABLE
 - Static variable in class org.apache.hadoop.hbase.coprocessor.example.TestZooKeeperScanPolicyObserver
+ 
 table
 - Variable in class org.apache.hadoop.hbase.coprocessor.TestHTableWrapper
  
 table
 - Variabl

[40/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.ZKWatcher.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.ZKWatcher.html
 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.ZKWatcher.html
deleted file mode 100644
index 25d3a06..000
--- 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.ZKWatcher.html
+++ /dev/null
@@ -1,402 +0,0 @@
-http://www.w3.org/TR/html4/loose.dtd";>
-
-
-
-
-
-ZooKeeperScanPolicyObserver.ZKWatcher (Apache HBase 3.0.0-SNAPSHOT 
API)
-
-
-
-
-
-var methods = {"i0":10,"i1":10};
-var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"]};
-var altColor = "altColor";
-var rowColor = "rowColor";
-var tableTab = "tableTab";
-var activeTableTab = "activeTableTab";
-
-
-JavaScript is disabled on your browser.
-
-
-
-
-
-Skip navigation links
-
-
-
-
-Overview
-Package
-Class
-Use
-Tree
-Deprecated
-Index
-Help
-
-
-
-
-Prev Class
-Next Class
-
-
-Frames
-No Frames
-
-
-All Classes
-
-
-
-
-
-
-
-Summary: 
-Nested | 
-Field | 
-Constr | 
-Method
-
-
-Detail: 
-Field | 
-Constr | 
-Method
-
-
-
-
-
-
-
-
-org.apache.hadoop.hbase.coprocessor.example
-Class 
ZooKeeperScanPolicyObserver.ZKWatcher
-
-
-
-http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">java.lang.Object
-
-
-org.apache.hadoop.hbase.coprocessor.example.ZooKeeperScanPolicyObserver.ZKWatcher
-
-
-
-
-
-
-
-All Implemented Interfaces:
-org.apache.zookeeper.Watcher
-
-
-Enclosing class:
-ZooKeeperScanPolicyObserver
-
-
-
-private static class ZooKeeperScanPolicyObserver.ZKWatcher
-extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
-implements org.apache.zookeeper.Watcher
-Internal watcher that keep "data" up to date 
asynchronously.
-
-
-
-
-
-
-
-
-
-
-
-Nested Class Summary
-
-
-
-
-Nested classes/interfaces inherited from 
interface org.apache.zookeeper.Watcher
-org.apache.zookeeper.Watcher.Event
-
-
-
-
-
-
-
-
-Field Summary
-
-Fields 
-
-Modifier and Type
-Field and Description
-
-
-private byte[]
-data 
-
-
-private long
-lastSetupTry 
-
-
-private boolean
-needSetup 
-
-
-private org.apache.zookeeper.ZooKeeper
-zk 
-
-
-
-
-
-
-
-
-
-Constructor Summary
-
-Constructors 
-
-Constructor and Description
-
-
-ZKWatcher(org.apache.zookeeper.ZooKeeper zk) 
-
-
-
-
-
-
-
-
-
-Method Summary
-
-All Methods Instance Methods Concrete Methods 
-
-Modifier and Type
-Method and Description
-
-
-byte[]
-getData()
-Get the maintained data.
-
-
-
-void
-process(org.apache.zookeeper.WatchedEvent event) 
-
-
-
-
-
-
-Methods inherited from class java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
-http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#clone--";
 title="class or interface in java.lang">clone, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#equals-java.lang.Object-";
 title="class or interface in java.lang">equals, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#finalize--";
 title="class or interface in java.lang">finalize, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#getClass--";
 title="class or interface in java.lang">getClass, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#hashCode--";
 title="class or interface in java.lang">hashCode, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#notify--";
 title="class or interface in java.lang">notify, http://docs.oracle.com/javase/8/docs/api/java/lang
 /Object.html?is-external=true#notifyAll--" title="class or interface in 
java.lang">notifyAll, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--";
 title="class or interface in java.lang">toString, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait--";
 title="class or interface in java.lang">wait, http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#wait-long-";
 title="class or interface in java.lang">wait, http://docs.oracle.com/java

[25/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/KeyValueHeap.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/KeyValueHeap.html 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/KeyValueHeap.html
index 35716ee..258f095 100644
--- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/KeyValueHeap.html
+++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/KeyValueHeap.html
@@ -148,317 +148,311 @@
 140   * 

141 * This can ONLY be called when you are using Scanners that implement InternalScanner as well as 142 * KeyValueScanner (a {@link StoreScanner}). -143 * @param result -144 * @return true if more rows exist after this one, false if scanner is done -145 */ -146 @Override -147 public boolean next(List result) throws IOException { -148return next(result, NoLimitScannerContext.getInstance()); -149 } -150 -151 @Override -152 public boolean next(List result, ScannerContext scannerContext) throws IOException { -153if (this.current == null) { -154 return scannerContext.setScannerState(NextState.NO_MORE_VALUES).hasMoreValues(); -155} -156InternalScanner currentAsInternal = (InternalScanner)this.current; -157boolean moreCells = currentAsInternal.next(result, scannerContext); -158Cell pee = this.current.peek(); -159 -160/* -161 * By definition, any InternalScanner must return false only when it has no -162 * further rows to be fetched. So, we can close a scanner if it returns -163 * false. All existing implementations seem to be fine with this. It is much -164 * more efficient to close scanners which are not needed than keep them in -165 * the heap. This is also required for certain optimizations. -166 */ -167 -168if (pee == null || !moreCells) { -169 // add the scanner that is to be closed -170 this.scannersForDelayedClose.add(this.current); -171} else { -172 this.heap.add(this.current); -173} -174this.current = null; -175this.current = pollRealKV(); -176if (this.current == null) { -177 moreCells = scannerContext.setScannerState(NextState.NO_MORE_VALUES).hasMoreValues(); -178} -179return moreCells; -180 } -181 -182 protected static class KVScannerComparator implements Comparator { -183protected CellComparator kvComparator; -184/** -185 * Constructor -186 * @param kvComparator -187 */ -188public KVScannerComparator(CellComparator kvComparator) { -189 this.kvComparator = kvComparator; -190} -191public int compare(KeyValueScanner left, KeyValueScanner right) { -192 int comparison = compare(left.peek(), right.peek()); -193 if (comparison != 0) { -194return comparison; -195 } else { -196// Since both the keys are exactly the same, we break the tie in favor of higher ordered -197// scanner since it'll have newer data. Since higher value should come first, we reverse -198// sort here. -199return Long.compare(right.getScannerOrder(), left.getScannerOrder()); -200 } -201} -202/** -203 * Compares two KeyValue -204 * @param left -205 * @param right -206 * @return less than 0 if left is smaller, 0 if equal etc.. +143 * @return true if more rows exist after this one, false if scanner is done +144 */ +145 @Override +146 public boolean next(List result, ScannerContext scannerContext) throws IOException { +147if (this.current == null) { +148 return scannerContext.setScannerState(NextState.NO_MORE_VALUES).hasMoreValues(); +149} +150InternalScanner currentAsInternal = (InternalScanner)this.current; +151boolean moreCells = currentAsInternal.next(result, scannerContext); +152Cell pee = this.current.peek(); +153 +154/* +155 * By definition, any InternalScanner must return false only when it has no +156 * further rows to be fetched. So, we can close a scanner if it returns +157 * false. All existing implementations seem to be fine with this. It is much +158 * more efficient to close scanners which are not needed than keep them in +159 * the heap. This is also required for certain optimizations. +160 */ +161 +162if (pee == null || !moreCells) { +163 // add the scanner that is to be closed +164 this.scannersForDelayedClose.add(this.current); +165} else { +166 this.heap.add(this.current); +167} +168this.current = null; +169this.current = pollRealKV(); +170if (this.current == null) { +171 moreCells = scannerContext.setScannerState(NextState.NO_MORE_VALUES).hasMoreValues(); +172} +173return moreCells; +174 } +175 +176 protected static class KVScannerComparator implements Comparator { +177protected


[17/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-build-configuration/hbase-archetypes/dependency-info.html
--
diff --git a/hbase-build-configuration/hbase-archetypes/dependency-info.html 
b/hbase-build-configuration/hbase-archetypes/dependency-info.html
index 21d9cc3..6f2c35a 100644
--- a/hbase-build-configuration/hbase-archetypes/dependency-info.html
+++ b/hbase-build-configuration/hbase-archetypes/dependency-info.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Archetypes – Dependency Information
 
@@ -148,7 +148,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-10-13
+  Last Published: 
2017-10-14
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-build-configuration/hbase-archetypes/dependency-management.html
--
diff --git 
a/hbase-build-configuration/hbase-archetypes/dependency-management.html 
b/hbase-build-configuration/hbase-archetypes/dependency-management.html
index 3d6973f..536a639 100644
--- a/hbase-build-configuration/hbase-archetypes/dependency-management.html
+++ b/hbase-build-configuration/hbase-archetypes/dependency-management.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Archetypes – Project Dependency 
Management
 
@@ -243,384 +243,390 @@
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>The Apache Software 
License, Version 2.0
 
+org.apache.curator
+http://curator.apache.org/curator-recipes";>curator-recipes
+4.0.0
+jar
+http://www.apache.org/licenses/LICENSE-2.0.txt";>The Apache Software 
License, Version 2.0
+
 org.apache.hadoop
 hadoop-auth
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-client
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-common
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-hdfs
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-mapreduce-client-core
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-mapreduce-client-jobclient
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-minicluster
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-annotations";>hbase-annotations
 3.0.0-SNAPSHOT
 test-jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-client";>hbase-client
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-common";>hbase-common
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-endpoint";>hbase-endpoint
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-support/hbase-error-prone";>hbase-error-prone
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-examples";>hbase-examples
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-external-blockcache";>hbase-external-blockcache
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-hadoop-compat";>hbase-hadoop-compat
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-hadoop2-compat";>hbase-hadoop2-compat
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-mapreduce";>hbase-mapreduce
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-metrics"

[32/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupCopyJob.SnapshotCopy.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupCopyJob.SnapshotCopy.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupCopyJob.SnapshotCopy.html
index 65a31ce..7551292 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupCopyJob.SnapshotCopy.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupCopyJob.SnapshotCopy.html
@@ -36,333 +36,414 @@
 028import org.apache.commons.logging.Log;
 029import 
org.apache.commons.logging.LogFactory;
 030import 
org.apache.hadoop.conf.Configuration;
-031import org.apache.hadoop.fs.FileSystem;
-032import org.apache.hadoop.fs.Path;
-033import 
org.apache.hadoop.hbase.TableName;
-034import 
org.apache.hadoop.hbase.backup.BackupCopyJob;
-035import 
org.apache.hadoop.hbase.backup.BackupInfo;
-036import 
org.apache.hadoop.hbase.backup.BackupType;
-037import 
org.apache.hadoop.hbase.backup.impl.BackupManager;
-038import 
org.apache.hadoop.hbase.backup.util.BackupUtils;
-039import 
org.apache.hadoop.hbase.snapshot.ExportSnapshot;
-040import 
org.apache.hadoop.mapreduce.Cluster;
-041import 
org.apache.hadoop.mapreduce.Counters;
-042import org.apache.hadoop.mapreduce.Job;
-043import 
org.apache.hadoop.mapreduce.JobID;
-044import org.apache.hadoop.tools.DistCp;
-045import 
org.apache.hadoop.tools.DistCpConstants;
-046import 
org.apache.hadoop.tools.DistCpOptions;
-047import 
org.apache.yetus.audience.InterfaceAudience;
-048import 
org.apache.zookeeper.KeeperException.NoNodeException;
-049
-050/**
-051 * Map-Reduce implementation of {@link 
BackupCopyJob}. Basically, there are 2 types of copy
-052 * operation: one is copying from 
snapshot, which bases on extending ExportSnapshot's function, the
-053 * other is copying for incremental log 
files, which bases on extending DistCp's function.
-054 */
-055@InterfaceAudience.Private
-056public class MapReduceBackupCopyJob 
implements BackupCopyJob {
-057  private static final Log LOG = 
LogFactory.getLog(MapReduceBackupCopyJob.class);
-058
-059  private Configuration conf;
-060
-061  // Accumulated progress within the 
whole backup process for the copy operation
-062  private float progressDone = 0.1f;
-063  private long bytesCopied = 0;
-064  private static float INIT_PROGRESS = 
0.1f;
+031import org.apache.hadoop.fs.FileStatus;
+032import org.apache.hadoop.fs.FileSystem;
+033import org.apache.hadoop.fs.Path;
+034import 
org.apache.hadoop.hbase.TableName;
+035import 
org.apache.hadoop.hbase.backup.BackupCopyJob;
+036import 
org.apache.hadoop.hbase.backup.BackupInfo;
+037import 
org.apache.hadoop.hbase.backup.BackupType;
+038import 
org.apache.hadoop.hbase.backup.impl.BackupManager;
+039import 
org.apache.hadoop.hbase.backup.util.BackupUtils;
+040import 
org.apache.hadoop.hbase.snapshot.ExportSnapshot;
+041import 
org.apache.hadoop.io.SequenceFile;
+042import org.apache.hadoop.io.Text;
+043import 
org.apache.hadoop.mapreduce.Cluster;
+044import 
org.apache.hadoop.mapreduce.Counters;
+045import org.apache.hadoop.mapreduce.Job;
+046import 
org.apache.hadoop.mapreduce.JobID;
+047import 
org.apache.hadoop.tools.CopyListingFileStatus;
+048import org.apache.hadoop.tools.DistCp;
+049import 
org.apache.hadoop.tools.DistCpConstants;
+050import 
org.apache.hadoop.tools.DistCpOptions;
+051import 
org.apache.yetus.audience.InterfaceAudience;
+052import 
org.apache.zookeeper.KeeperException.NoNodeException;
+053
+054/**
+055 * Map-Reduce implementation of {@link 
BackupCopyJob}. Basically, there are 2 types of copy
+056 * operation: one is copying from 
snapshot, which bases on extending ExportSnapshot's function, the
+057 * other is copying for incremental log 
files, which bases on extending DistCp's function.
+058 */
+059@InterfaceAudience.Private
+060public class MapReduceBackupCopyJob 
implements BackupCopyJob {
+061  public static final String 
NUMBER_OF_LEVELS_TO_PRESERVE_KEY = "num.levels.preserve";
+062  private static final Log LOG = 
LogFactory.getLog(MapReduceBackupCopyJob.class);
+063
+064  private Configuration conf;
 065
-066  // The percentage of the current copy 
task within the whole task if multiple time copies are
-067  // needed. The default value is 100%, 
which means only 1 copy task for the whole.
-068  private float subTaskPercntgInWholeTask 
= 1f;
-069
-070  public MapReduceBackupCopyJob() {
-071  }
-072
-073  @Override
-074  public Configuration getConf() {
-075return conf;
+066  // Accumulated progress within the 
whole backup process for the copy operation
+067  private float progressDone = 0.1f;
+068  private long bytesCopied = 0;
+069  private static float INIT_PROGRESS = 
0.1f;
+070
+071  // The percentage of the current copy 
task within 

[03/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/testdevapidocs/src-html/org/apache/hadoop/hbase/backup/TestBackupBase.FullTableBackupClientForTest.html
--
diff --git 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/backup/TestBackupBase.FullTableBackupClientForTest.html
 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/backup/TestBackupBase.FullTableBackupClientForTest.html
index 11a0572..4c6fde8 100644
--- 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/backup/TestBackupBase.FullTableBackupClientForTest.html
+++ 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/backup/TestBackupBase.FullTableBackupClientForTest.html
@@ -144,375 +144,376 @@
 136// copy out the table and region 
info files for each table
 137
BackupUtils.copyTableRegionInfo(conn, backupInfo, conf);
 138// convert WAL to HFiles and copy 
them to .tmp under BACKUP_ROOT
-139
convertWALsToHFiles(backupInfo);
-140
incrementalCopyHFiles(backupInfo);
-141failStageIf(Stage.stage_2);
-142// Save list of WAL files 
copied
-143
backupManager.recordWALFiles(backupInfo.getIncrBackupFileList());
-144
-145// case INCR_BACKUP_COMPLETE:
-146// set overall backup status: 
complete. Here we make sure to complete the backup.
-147// After this checkpoint, even if 
entering cancel process, will let the backup finished
-148// Set the previousTimestampMap 
which is before this current log roll to the manifest.
-149HashMap> previousTimestampMap =
-150
backupManager.readLogTimestampMap();
-151
backupInfo.setIncrTimestampMap(previousTimestampMap);
-152
-153// The table list in backupInfo 
is good for both full backup and incremental backup.
-154// For incremental backup, it 
contains the incremental backup table set.
-155
backupManager.writeRegionServerLogTimestamp(backupInfo.getTables(), 
newTimestamps);
-156failStageIf(Stage.stage_3);
-157
-158HashMap> newTableSetTimestampMap =
-159
backupManager.readLogTimestampMap();
-160
-161Long newStartCode =
-162
BackupUtils.getMinValue(BackupUtils.getRSLogTimestampMins(newTableSetTimestampMap));
-163
backupManager.writeBackupStartCode(newStartCode);
-164
-165
handleBulkLoad(backupInfo.getTableNames());
-166failStageIf(Stage.stage_4);
-167
-168// backup complete
-169completeBackup(conn, backupInfo, 
backupManager, BackupType.INCREMENTAL, conf);
-170
-171  } catch (Exception e) {
-172failBackup(conn, backupInfo, 
backupManager, e, "Unexpected Exception : ",
-173  BackupType.INCREMENTAL, 
conf);
-174throw new IOException(e);
-175  }
-176
-177}
-178  }
-179
-180  static class 
FullTableBackupClientForTest extends FullTableBackupClient
-181  {
-182
+139convertWALsToHFiles();
+140incrementalCopyHFiles(new 
String[] {getBulkOutputDir().toString()},
+141  
backupInfo.getBackupRootDir());
+142failStageIf(Stage.stage_2);
+143// Save list of WAL files 
copied
+144
backupManager.recordWALFiles(backupInfo.getIncrBackupFileList());
+145
+146// case INCR_BACKUP_COMPLETE:
+147// set overall backup status: 
complete. Here we make sure to complete the backup.
+148// After this checkpoint, even if 
entering cancel process, will let the backup finished
+149// Set the previousTimestampMap 
which is before this current log roll to the manifest.
+150HashMap> previousTimestampMap =
+151
backupManager.readLogTimestampMap();
+152
backupInfo.setIncrTimestampMap(previousTimestampMap);
+153
+154// The table list in backupInfo 
is good for both full backup and incremental backup.
+155// For incremental backup, it 
contains the incremental backup table set.
+156
backupManager.writeRegionServerLogTimestamp(backupInfo.getTables(), 
newTimestamps);
+157failStageIf(Stage.stage_3);
+158
+159HashMap> newTableSetTimestampMap =
+160
backupManager.readLogTimestampMap();
+161
+162Long newStartCode =
+163
BackupUtils.getMinValue(BackupUtils.getRSLogTimestampMins(newTableSetTimestampMap));
+164
backupManager.writeBackupStartCode(newStartCode);
+165
+166
handleBulkLoad(backupInfo.getTableNames());
+167failStageIf(Stage.stage_4);
+168
+169// backup complete
+170completeBackup(conn, backupInfo, 
backupManager, BackupType.INCREMENTAL, conf);
+171
+172  } catch (Exception e) {
+173failBackup(conn, backupInfo, 
backupManager, e, "Unexpected Exception : ",
+174  BackupType.INCREMENTAL, 
conf);
+175throw new IOException(e);
+176  }
+177

[49/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/apidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.html
--
diff --git a/apidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.html 
b/apidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.html
index e48f3fd..fcfe8ab 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/FilterList.html
@@ -29,536 +29,554 @@
 021import java.io.IOException;
 022import java.util.ArrayList;
 023import java.util.Arrays;
-024import java.util.Collections;
-025import java.util.List;
-026
-027import org.apache.hadoop.hbase.Cell;
-028import 
org.apache.hadoop.hbase.CellComparator;
-029import 
org.apache.hadoop.hbase.CellUtil;
-030import 
org.apache.yetus.audience.InterfaceAudience;
-031import 
org.apache.hadoop.hbase.exceptions.DeserializationException;
-032import 
org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
-033import 
org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos;
-034import 
org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException;
-035
-036/**
-037 * Implementation of {@link Filter} that 
represents an ordered List of Filters
-038 * which will be evaluated with a 
specified boolean operator {@link Operator#MUST_PASS_ALL}
-039 * (AND) or 
{@link Operator#MUST_PASS_ONE} (OR).
-040 * Since you can use Filter Lists as 
children of Filter Lists, you can create a
-041 * hierarchy of filters to be 
evaluated.
-042 *
-043 * 
-044 * {@link Operator#MUST_PASS_ALL} evaluates lazily: evaluation stops as soon as one filter does -045 * not include the KeyValue. -046 * -047 *
-048 * {@link Operator#MUST_PASS_ONE} evaluates non-lazily: all filters are always evaluated. -049 * -050 *
-051 * Defaults to {@link Operator#MUST_PASS_ALL}. -052 */ -053@InterfaceAudience.Public -054final public class FilterList extends FilterBase { -055 /** set operator */ -056 @InterfaceAudience.Public -057 public static enum Operator { -058/** !AND */ -059MUST_PASS_ALL, -060/** !OR */ -061MUST_PASS_ONE -062 } -063 -064 private static final int MAX_LOG_FILTERS = 5; -065 private Operator operator = Operator.MUST_PASS_ALL; -066 private final List filters; -067 private Filter seekHintFilter = null; -068 -069 /** Reference Cell used by {@link #transformCell(Cell)} for validation purpose. */ -070 private Cell referenceCell = null; -071 -072 /** -073 * When filtering a given Cell in {@link #filterKeyValue(Cell)}, -074 * this stores the transformed Cell to be returned by {@link #transformCell(Cell)}. -075 * -076 * Individual filters transformation are applied only when the filter includes the Cell. -077 * Transformations are composed in the order specified by {@link #filters}. -078 */ -079 private Cell transformedCell = null; -080 -081 /** -082 * Constructor that takes a set of {@link Filter}s. The default operator -083 * MUST_PASS_ALL is assumed. -084 * All filters are cloned to internal list. -085 * @param rowFilters list of filters -086 */ -087 public FilterList(final List rowFilters) { -088reversed = getReversed(rowFilters, reversed); -089this.filters = new ArrayList<>(rowFilters); -090 } -091 -092 /** -093 * Constructor that takes a var arg number of {@link Filter}s. The fefault operator -094 * MUST_PASS_ALL is assumed. -095 * @param rowFilters -096 */ -097 public FilterList(final Filter... rowFilters) { -098this(Arrays.asList(rowFilters)); -099 } -100 -101 /** -102 * Constructor that takes an operator. -103 * -104 * @param operator Operator to process filter set with. -105 */ -106 public FilterList(final Operator operator) { -107this.operator = operator; -108this.filters = new ArrayList<>(); -109 } -110 -111 /** -112 * Constructor that takes a set of {@link Filter}s and an operator. -113 * -114 * @param operator Operator to process filter set with. -115 * @param rowFilters Set of row filters. -116 */ -117 public FilterList(final Operator operator, final List rowFilters) { -118this(rowFilters); -119this.operator = operator; -120 } -121 -122 /** -123 * Constructor that takes a var arg number of {@link Filter}s and an operator. -124 * -125 * @param operator Operator to process filter set with. -126 * @param rowFilters Filters to use -127 */ -128 public FilterList(final Operator operator, final Filter... rowFilters) { -129this(rowFilters); -130this.operator = operator; -131 } -132 -133 /** -134 * Get the operator. -135 * -136 * @return operator -137 */ -138 public Operator getOperator() { -139return operator; -140 } -141 -142 /** -143 * Get the filters. -144 * -145 * @return filters -146 */ -147 public List getFilters() { -148return fil

[37/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
index d4258d1..8a60264 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/package-tree.html
@@ -320,11 +320,11 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.master.RegionState.State
-org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus
 org.apache.hadoop.hbase.master.MasterRpcServices.BalanceSwitchMode
 org.apache.hadoop.hbase.master.MetricsMasterSourceFactoryImpl.FactoryStorage
+org.apache.hadoop.hbase.master.RegionState.State
 org.apache.hadoop.hbase.master.SplitLogManager.ResubmitDirective
+org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html
index 509ef50..a6a1367 100644
--- a/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/master/procedure/package-tree.html
@@ -212,8 +212,8 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.master.procedure.TableProcedureInterface.TableOperationType
 org.apache.hadoop.hbase.master.procedure.ServerProcedureInterface.ServerOperationType
+org.apache.hadoop.hbase.master.procedure.TableProcedureInterface.TableOperationType
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/org/apache/hadoop/hbase/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/package-tree.html
index 028af1e..e2c4b2a 100644
--- a/devapidocs/org/apache/hadoop/hbase/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/package-tree.html
@@ -427,17 +427,17 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.Coprocessor.State
 org.apache.hadoop.hbase.KeyValue.Type
-org.apache.hadoop.hbase.ProcedureState
 org.apache.hadoop.hbase.CellBuilder.DataType
-org.apache.hadoop.hbase.CellBuilderType
 org.apache.hadoop.hbase.HealthChecker.HealthCheckerExitStatus
-org.apache.hadoop.hbase.ClusterStatus.Option
-org.apache.hadoop.hbase.HConstants.OperationStatusCode
 org.apache.hadoop.hbase.KeepDeletedCells
-org.apache.hadoop.hbase.CompatibilitySingletonFactory.SingletonStorage
 org.apache.hadoop.hbase.CompareOperator
+org.apache.hadoop.hbase.ProcedureState
+org.apache.hadoop.hbase.CellBuilderType
+org.apache.hadoop.hbase.HConstants.OperationStatusCode
+org.apache.hadoop.hbase.ClusterStatus.Option
+org.apache.hadoop.hbase.CompatibilitySingletonFactory.SingletonStorage
+org.apache.hadoop.hbase.Coprocessor.State
 org.apache.hadoop.hbase.MemoryCompactionPolicy
 org.apache.hadoop.hbase.MetaTableAccessor.QueryType
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html
index 4277290..50ad8c3 100644
--- a/devapidocs/org/apache/hadoop/hbase/procedure2/pack

[02/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/testdevapidocs/src-html/org/apache/hadoop/hbase/backup/TestBackupBase.IncrementalTableBackupClientForTest.html
--
diff --git 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/backup/TestBackupBase.IncrementalTableBackupClientForTest.html
 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/backup/TestBackupBase.IncrementalTableBackupClientForTest.html
index 11a0572..4c6fde8 100644
--- 
a/testdevapidocs/src-html/org/apache/hadoop/hbase/backup/TestBackupBase.IncrementalTableBackupClientForTest.html
+++ 
b/testdevapidocs/src-html/org/apache/hadoop/hbase/backup/TestBackupBase.IncrementalTableBackupClientForTest.html
@@ -144,375 +144,376 @@
 136// copy out the table and region 
info files for each table
 137
BackupUtils.copyTableRegionInfo(conn, backupInfo, conf);
 138// convert WAL to HFiles and copy 
them to .tmp under BACKUP_ROOT
-139
convertWALsToHFiles(backupInfo);
-140
incrementalCopyHFiles(backupInfo);
-141failStageIf(Stage.stage_2);
-142// Save list of WAL files 
copied
-143
backupManager.recordWALFiles(backupInfo.getIncrBackupFileList());
-144
-145// case INCR_BACKUP_COMPLETE:
-146// set overall backup status: 
complete. Here we make sure to complete the backup.
-147// After this checkpoint, even if 
entering cancel process, will let the backup finished
-148// Set the previousTimestampMap 
which is before this current log roll to the manifest.
-149HashMap> previousTimestampMap =
-150
backupManager.readLogTimestampMap();
-151
backupInfo.setIncrTimestampMap(previousTimestampMap);
-152
-153// The table list in backupInfo 
is good for both full backup and incremental backup.
-154// For incremental backup, it 
contains the incremental backup table set.
-155
backupManager.writeRegionServerLogTimestamp(backupInfo.getTables(), 
newTimestamps);
-156failStageIf(Stage.stage_3);
-157
-158HashMap> newTableSetTimestampMap =
-159
backupManager.readLogTimestampMap();
-160
-161Long newStartCode =
-162
BackupUtils.getMinValue(BackupUtils.getRSLogTimestampMins(newTableSetTimestampMap));
-163
backupManager.writeBackupStartCode(newStartCode);
-164
-165
handleBulkLoad(backupInfo.getTableNames());
-166failStageIf(Stage.stage_4);
-167
-168// backup complete
-169completeBackup(conn, backupInfo, 
backupManager, BackupType.INCREMENTAL, conf);
-170
-171  } catch (Exception e) {
-172failBackup(conn, backupInfo, 
backupManager, e, "Unexpected Exception : ",
-173  BackupType.INCREMENTAL, 
conf);
-174throw new IOException(e);
-175  }
-176
-177}
-178  }
-179
-180  static class 
FullTableBackupClientForTest extends FullTableBackupClient
-181  {
-182
+139convertWALsToHFiles();
+140incrementalCopyHFiles(new 
String[] {getBulkOutputDir().toString()},
+141  
backupInfo.getBackupRootDir());
+142failStageIf(Stage.stage_2);
+143// Save list of WAL files 
copied
+144
backupManager.recordWALFiles(backupInfo.getIncrBackupFileList());
+145
+146// case INCR_BACKUP_COMPLETE:
+147// set overall backup status: 
complete. Here we make sure to complete the backup.
+148// After this checkpoint, even if 
entering cancel process, will let the backup finished
+149// Set the previousTimestampMap 
which is before this current log roll to the manifest.
+150HashMap> previousTimestampMap =
+151
backupManager.readLogTimestampMap();
+152
backupInfo.setIncrTimestampMap(previousTimestampMap);
+153
+154// The table list in backupInfo 
is good for both full backup and incremental backup.
+155// For incremental backup, it 
contains the incremental backup table set.
+156
backupManager.writeRegionServerLogTimestamp(backupInfo.getTables(), 
newTimestamps);
+157failStageIf(Stage.stage_3);
+158
+159HashMap> newTableSetTimestampMap =
+160
backupManager.readLogTimestampMap();
+161
+162Long newStartCode =
+163
BackupUtils.getMinValue(BackupUtils.getRSLogTimestampMins(newTableSetTimestampMap));
+164
backupManager.writeBackupStartCode(newStartCode);
+165
+166
handleBulkLoad(backupInfo.getTableNames());
+167failStageIf(Stage.stage_4);
+168
+169// backup complete
+170completeBackup(conn, backupInfo, 
backupManager, BackupType.INCREMENTAL, conf);
+171
+172  } catch (Exception e) {
+173failBackup(conn, backupInfo, 
backupManager, e, "Unexpected Exception : ",
+174  BackupType.INCREMENTAL, 
conf);
+175throw ne

[27/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
index eac35d3..5d138ea 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html
@@ -7287,10 +7287,10 @@
 7279  try {
 7280processor.process(now, region, 
mutations, walEdit);
 7281  } catch (IOException e) {
-7282LOG.warn("RowProcessor:" + 
processor.getClass().getName() +
-7283" throws Exception on 
row(s):" +
-7284Bytes.toStringBinary(
-7285  
processor.getRowsToLock().iterator().next()) + "...", e);
+7282String row = 
processor.getRowsToLock().isEmpty() ? "" :
+7283  " on row(s):" + 
Bytes.toStringBinary(processor.getRowsToLock().iterator().next()) + "...";
+7284LOG.warn("RowProcessor:" + 
processor.getClass().getName() +
+7285" throws Exception" + row, 
e);
 7286throw e;
 7287  }
 7288  return;
@@ -7304,10 +7304,10 @@
 7296processor.process(now, 
region, mutations, walEdit);
 7297return null;
 7298  } catch (IOException e) {
-7299LOG.warn("RowProcessor:" + 
processor.getClass().getName() +
-7300" throws Exception on 
row(s):" +
-7301Bytes.toStringBinary(
-7302
processor.getRowsToLock().iterator().next()) + "...", e);
+7299String row = 
processor.getRowsToLock().isEmpty() ? "" :
+7300  " on row(s):" + 
Bytes.toStringBinary(processor.getRowsToLock().iterator().next()) + "...";
+7301LOG.warn("RowProcessor:" + 
processor.getClass().getName() +
+7302" throws Exception" + 
row, e);
 7303throw e;
 7304  }
 7305}
@@ -7316,9 +7316,9 @@
 7308try {
 7309  task.get(timeout, 
TimeUnit.MILLISECONDS);
 7310} catch (TimeoutException te) {
-7311  LOG.error("RowProcessor timeout:" 
+ timeout + " ms on row(s):" +
-7312  
Bytes.toStringBinary(processor.getRowsToLock().iterator().next()) +
-7313  "...");
+7311  String row = 
processor.getRowsToLock().isEmpty() ? "" :
+7312" on row(s):" + 
Bytes.toStringBinary(processor.getRowsToLock().iterator().next()) + "...";
+7313  LOG.error("RowProcessor timeout:" 
+ timeout + " ms" + row);
 7314  throw new IOException(te);
 7315} catch (Exception e) {
 7316  throw new IOException(e);

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
index eac35d3..5d138ea 100644
--- 
a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html
@@ -7287,10 +7287,10 @@
 7279  try {
 7280processor.process(now, region, 
mutations, walEdit);
 7281  } catch (IOException e) {
-7282LOG.warn("RowProcessor:" + 
processor.getClass().getName() +
-7283" throws Exception on 
row(s):" +
-7284Bytes.toStringBinary(
-7285  
processor.getRowsToLock().iterator().next()) + "...", e);
+7282String row = 
processor.getRowsToLock().isEmpty() ? "" :
+7283  " on row(s):" + 
Bytes.toStringBinary(processor.getRowsToLock().iterator().next()) + "...";
+7284LOG.warn("RowProcessor:" + 
processor.getClass().getName() +
+7285" throws Exception" + row, 
e);
 7286throw e;
 7287  }
 7288  return;
@@ -7304,10 +7304,10 @@
 7296processor.process(now, 
region, mutations, walEdit);
 7297return null;
 7298  } catch (IOException e) {
-7299LOG.warn("RowProcessor:" + 
processor.getClass().getName() +
-7300" throws Exception on 
row(s):" +
-7301Bytes.toStringBinary(
-7302
processor.getRowsToLock().iterator().next()) + "...", e);
+7299String row = 
processor.getRowsToLock().isEmpty() ? "" :
+7300  " on row(s):" + 
Bytes.toStringBinary(processor.getRowsToLock().iterator().next()) + "...";
+7301LOG.warn("RowProcessor:" + 
processor.getClass().getName() +
+7302" throws Exception" + 
row, e);
 7303throw e;
 7304  }
 7305   

[13/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
--
diff --git 
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
 
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
index 28dfe9f..96f033a 100644
--- 
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
+++ 
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-info.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Exemplar for hbase-client archetype – 
Dependency Information
 
@@ -147,7 +147,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-10-13
+  Last Published: 
2017-10-14
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
--
diff --git 
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
 
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
index 0ee53ac..fe9e07a 100644
--- 
a/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
+++ 
b/hbase-build-configuration/hbase-archetypes/hbase-client-project/dependency-management.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Exemplar for hbase-client archetype – Project 
Dependency Management
 
@@ -243,384 +243,390 @@
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>The Apache Software 
License, Version 2.0
 
+org.apache.curator
+http://curator.apache.org/curator-recipes";>curator-recipes
+4.0.0
+jar
+http://www.apache.org/licenses/LICENSE-2.0.txt";>The Apache Software 
License, Version 2.0
+
 org.apache.hadoop
 hadoop-auth
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-client
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-common
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-hdfs
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-mapreduce-client-core
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-mapreduce-client-jobclient
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-minicluster
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-annotations";>hbase-annotations
 3.0.0-SNAPSHOT
 test-jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-client";>hbase-client
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-common";>hbase-common
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-endpoint";>hbase-endpoint
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-support/hbase-error-prone";>hbase-error-prone
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-examples";>hbase-examples
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-external-blockcache";>hbase-external-blockcache
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-hadoop-compat";>hbase-hadoop-compat
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-hadoop2-compat";>hbase-hadoop2-compat
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http

[30/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.ZKDataHolder.html
--
diff --git 
a/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.ZKDataHolder.html
 
b/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.ZKDataHolder.html
new file mode 100644
index 000..eef9d8d
--- /dev/null
+++ 
b/devapidocs/src-html/org/apache/hadoop/hbase/coprocessor/example/ZooKeeperScanPolicyObserver.ZKDataHolder.html
@@ -0,0 +1,273 @@
+http://www.w3.org/TR/html4/loose.dtd";>
+
+
+Source code
+
+
+
+
+001/**
+002 * Copyright The Apache Software 
Foundation
+003 *
+004 * Licensed to the Apache Software 
Foundation (ASF) under one or more
+005 * contributor license agreements. See 
the NOTICE file distributed with this
+006 * work for additional information 
regarding copyright ownership. The ASF
+007 * licenses this file to you under the 
Apache License, Version 2.0 (the
+008 * "License"); you may not use this file 
except in compliance with the License.
+009 * You may obtain a copy of the License 
at
+010 *
+011 * 
http://www.apache.org/licenses/LICENSE-2.0
+012 *
+013 * Unless required by applicable law or 
agreed to in writing, software
+014 * distributed under the License is 
distributed on an "AS IS" BASIS, WITHOUT
+015 * WARRANTIES OR CONDITIONS OF ANY KIND, 
either express or implied. See the
+016 * License for the specific language 
governing permissions and limitations
+017 * under the License.
+018 */
+019package 
org.apache.hadoop.hbase.coprocessor.example;
+020
+021import java.io.IOException;
+022import java.util.List;
+023import java.util.Optional;
+024import java.util.OptionalLong;
+025
+026import 
org.apache.curator.framework.CuratorFramework;
+027import 
org.apache.curator.framework.CuratorFrameworkFactory;
+028import 
org.apache.curator.framework.recipes.cache.ChildData;
+029import 
org.apache.curator.framework.recipes.cache.NodeCache;
+030import 
org.apache.curator.retry.RetryForever;
+031import org.apache.hadoop.hbase.Cell;
+032import 
org.apache.hadoop.hbase.CoprocessorEnvironment;
+033import 
org.apache.hadoop.hbase.coprocessor.ObserverContext;
+034import 
org.apache.hadoop.hbase.coprocessor.RegionCoprocessor;
+035import 
org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment;
+036import 
org.apache.hadoop.hbase.coprocessor.RegionObserver;
+037import 
org.apache.hadoop.hbase.regionserver.InternalScanner;
+038import 
org.apache.hadoop.hbase.regionserver.ScanType;
+039import 
org.apache.hadoop.hbase.regionserver.ScannerContext;
+040import 
org.apache.hadoop.hbase.regionserver.Store;
+041import 
org.apache.hadoop.hbase.regionserver.compactions.CompactionLifeCycleTracker;
+042import 
org.apache.hadoop.hbase.regionserver.compactions.CompactionRequest;
+043import 
org.apache.hadoop.hbase.util.Bytes;
+044
+045/**
+046 * This is an example showing how a 
RegionObserver could configured via ZooKeeper in order to
+047 * control a Region compaction, flush, 
and scan policy. This also demonstrated the use of shared
+048 * {@link 
org.apache.hadoop.hbase.coprocessor.RegionObserver} state. See
+049 * {@link 
RegionCoprocessorEnvironment#getSharedData()}.
+050 * 

+051 * This would be useful for an incremental backup tool, which would indicate the last time of a +052 * successful backup via ZK and instruct HBase that to safely delete the data which has already been +053 * backup. +054 */ +055public class ZooKeeperScanPolicyObserver implements RegionCoprocessor, RegionObserver { +056 +057 @Override +058 public Optional getRegionObserver() { +059return Optional.of(this); +060 } +061 +062 // The zk ensemble info is put in hbase config xml with given custom key. +063 public static final String ZK_ENSEMBLE_KEY = "ZooKeeperScanPolicyObserver.zookeeper.ensemble"; +064 public static final String ZK_SESSION_TIMEOUT_KEY = +065 "ZooKeeperScanPolicyObserver.zookeeper.session.timeout"; +066 public static final int ZK_SESSION_TIMEOUT_DEFAULT = 30 * 1000; // 30 secs +067 public static final String NODE = "/backup/example/lastbackup"; +068 private static final String ZKKEY = "ZK"; +069 +070 private NodeCache cache; +071 +072 /** +073 * Internal watcher that keep "data" up to date asynchronously. +074 */ +075 private static final class ZKDataHolder { +076 +077private final String ensemble; +078 +079private final int sessionTimeout; +080 +081private CuratorFramework client; +082 +083private NodeCache cache; +084 +085private int ref; +086 +087public ZKDataHolder(String ensemble, int sessionTimeout) { +088 this.ensemble = ensemble; +089 this.sessionTimeout = sessionTimeout; +090} +091 +092private void create() throws Exception { +093 client = +094 CuratorFramewo


[11/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-info.html
--
diff --git 
a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-info.html
 
b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-info.html
index 948bd15..fe3a24a 100644
--- 
a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-info.html
+++ 
b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-info.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Exemplar for hbase-shaded-client archetype – 
Dependency Information
 
@@ -147,7 +147,7 @@
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-10-13
+  Last Published: 
2017-10-14
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-management.html
--
diff --git 
a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-management.html
 
b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-management.html
index ce3f984..c3dfe27 100644
--- 
a/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-management.html
+++ 
b/hbase-build-configuration/hbase-archetypes/hbase-shaded-client-project/dependency-management.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase - Exemplar for hbase-shaded-client archetype – 
Project Dependency Management
 
@@ -243,384 +243,390 @@
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>The Apache Software 
License, Version 2.0
 
+org.apache.curator
+http://curator.apache.org/curator-recipes";>curator-recipes
+4.0.0
+jar
+http://www.apache.org/licenses/LICENSE-2.0.txt";>The Apache Software 
License, Version 2.0
+
 org.apache.hadoop
 hadoop-auth
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-client
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-common
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-hdfs
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-mapreduce-client-core
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-mapreduce-client-jobclient
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hadoop
 hadoop-minicluster
 2.7.1
 jar
 http://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-annotations";>hbase-annotations
 3.0.0-SNAPSHOT
 test-jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-client";>hbase-client
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-common";>hbase-common
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-endpoint";>hbase-endpoint
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-support/hbase-error-prone";>hbase-error-prone
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-examples";>hbase-examples
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-external-blockcache";>hbase-external-blockcache
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-hadoop-compat";>hbase-hadoop-compat
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org/licenses/LICENSE-2.0.txt";>Apache License, Version 
2.0
-
+
 org.apache.hbase
 http://hbase.apache.org/hbase-build-configuration/hbase-hadoop2-compat";>hbase-hadoop2-compat
 3.0.0-SNAPSHOT
 jar
 https://www.apache.org

[36/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/org/apache/hadoop/hbase/regionserver/StoreScanner.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/StoreScanner.html 
b/devapidocs/org/apache/hadoop/hbase/regionserver/StoreScanner.html
index 951c41e..2eeabe7 100644
--- a/devapidocs/org/apache/hadoop/hbase/regionserver/StoreScanner.html
+++ b/devapidocs/org/apache/hadoop/hbase/regionserver/StoreScanner.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":9,"i4":10,"i5":10,"i6":9,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":9,"i4":10,"i5":10,"i6":9,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -527,75 +527,69 @@ implements 
 boolean
-next(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List outResult)
-Grab the next row's worth of values.
-
-
-
-boolean
 next(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List outResult,
 ScannerContext scannerContext)
 Get the next row of values from this Store.
 
 
-
+
 private void
 parallelSeek(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List scanners,
 Cell kv)
 Seek storefiles in parallel to optimize IO latency as much 
as possible
 
 
-
+
 Cell
 peek()
 Look at the next Cell in this scanner, but do not iterate 
scanner.
 
 
-
+
 protected boolean
 reopenAfterFlush() 
 
-
+
 boolean
 reseek(Cell kv)
 Reseek the scanner at or after the specified KeyValue.
 
 
-
+
 protected void
 resetKVHeap(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List scanners,
CellComparator comparator) 
 
-
+
 private void
 resetQueryMatcher(Cell lastTopKey) 
 
-
+
 boolean
 seek(Cell key)
 Seek the scanner at or after the specified KeyValue.
 
 
-
+
 private void
 seekAllScanner(ScanInfo scanInfo,
   http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List scanners) 
 
-
+
 protected boolean
 seekAsDirection(Cell kv)
 Do a reseek in a normal StoreScanner(scan forward)
 
 
-
+
 private void
 seekOrSkipToNextColumn(Cell cell) 
 
-
+
 private void
 seekOrSkipToNextRow(Cell cell) 
 
-
+
 protected void
 seekScanners(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List scanners,
 Cell seekKey,
@@ -604,40 +598,40 @@ implements Seek the specified scanners with the given key
 
 
-
+
 protected boolean
 seekToNextRow(Cell c) 
 
-
+
 protected http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List
 selectScannersFrom(HStore store,
   http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List allScanners)
 Filters the given list of scanners using Bloom filter, time 
range, and TTL.
 
 
-
+
 void
 shipped()
 Called after a batch of rows scanned and set to be returned 
to client.
 
 
-
+
 protected boolean
 trySkipToNextColumn(Cell cell)
 See trySkipToNextRow(Cell)
 
 
-
+
 protected boolean
 trySkipToNextRow(Cell cell)
 See if we should actually SEEK or rather just SKIP to the 
next Cell (see HBASE-13109).
 
 
-
+
 (package private) void
 trySwitchToStreamRead() 
 
-
+
 void
 updateReaders(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List sfs,
  http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List memStoreScanners)
@@ -673,6 +667,13 @@ implements KeyValueScanner
 backwardSeek,
 enforceSeek,
 getFilePath,
 isFileScanner,
 realSeekDone,
 requestSeek,
 seekToLastRow,
 seekToPreviousRow, shouldUseScanner
 
+
+
+
+
+Methods inherited

[43/51] [partial] hbase-site git commit: Published site at .

2017-10-14 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/7816cbde/devapidocs/index-all.html
--
diff --git a/devapidocs/index-all.html b/devapidocs/index-all.html
index fe484d5..59c4126 100644
--- a/devapidocs/index-all.html
+++ b/devapidocs/index-all.html
@@ -688,6 +688,8 @@
 
 aclZNode
 - Variable in class org.apache.hadoop.hbase.security.access.ZKPermissionWatcher
  
+acquire()
 - Method in class org.apache.hadoop.hbase.coprocessor.example.ZooKeeperScanPolicyObserver.ZKDataHolder
+ 
 acquire()
 - Method in class org.apache.hadoop.hbase.master.locking.LockManager.MasterLock
 
 Acquire the lock, waiting indefinitely until the lock is 
released or
@@ -7803,6 +7805,10 @@
  
 cache 
- Variable in class org.apache.hadoop.hbase.client.ClientScanner
  
+cache
 - Variable in class org.apache.hadoop.hbase.coprocessor.example.ZooKeeperScanPolicyObserver
+ 
+cache
 - Variable in class org.apache.hadoop.hbase.coprocessor.example.ZooKeeperScanPolicyObserver.ZKDataHolder
+ 
 cache
 - Variable in class org.apache.hadoop.hbase.io.hfile.LruBlockCache.EvictionThread
  
 cache
 - Variable in class org.apache.hadoop.hbase.master.balancer.RegionLocationFinder
@@ -11940,6 +11946,8 @@
  
 ClickjackingPreventionFilter()
 - Constructor for class org.apache.hadoop.hbase.http.ClickjackingPreventionFilter
  
+client
 - Variable in class org.apache.hadoop.hbase.coprocessor.example.ZooKeeperScanPolicyObserver.ZKDataHolder
+ 
 client
 - Variable in class org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutput
  
 client
 - Variable in class org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputHelper.CancelOnClose
@@ -12326,6 +12334,10 @@
  
 close()
 - Method in class org.apache.hadoop.hbase.client.ZooKeeperKeepAliveConnection
  
+close()
 - Method in class org.apache.hadoop.hbase.coprocessor.example.DelegatingInternalScanner
+ 
+close()
 - Method in class org.apache.hadoop.hbase.coprocessor.example.ZooKeeperScanPolicyObserver.ZKDataHolder
+ 
 close()
 - Method in class org.apache.hadoop.hbase.coprocessor.Export.PrivilegedWriter
  
 close()
 - Method in class org.apache.hadoop.hbase.coprocessor.Export.RegionOp
@@ -17145,7 +17157,7 @@
  
 convertToV2SingleManifest()
 - Method in class org.apache.hadoop.hbase.snapshot.SnapshotManifest
  
-convertWALsToHFiles(BackupInfo)
 - Method in class org.apache.hadoop.hbase.backup.impl.IncrementalTableBackupClient
+convertWALsToHFiles()
 - Method in class org.apache.hadoop.hbase.backup.impl.IncrementalTableBackupClient
  
 coord 
- Variable in class org.apache.hadoop.hbase.procedure.Procedure
  
@@ -17512,6 +17524,8 @@
 
 Copy data from a buffer to an output stream.
 
+copyBulkLoadedFiles(List,
 List) - Method in class 
org.apache.hadoop.hbase.backup.impl.IncrementalTableBackupClient
+ 
 copyBytes()
 - Method in class org.apache.hadoop.hbase.io.ImmutableBytesWritable
 
 Returns a copy of the bytes referred to by this 
writable
@@ -18091,6 +18105,8 @@
  
 create(Configuration)
 - Static method in class org.apache.hadoop.hbase.client.ServerStatisticTracker
  
+create()
 - Method in class org.apache.hadoop.hbase.coprocessor.example.ZooKeeperScanPolicyObserver.ZKDataHolder
+ 
 create(CellBuilderType)
 - Static method in class org.apache.hadoop.hbase.ExtendedCellBuilderFactory
  
 create()
 - Static method in class org.apache.hadoop.hbase.HBaseConfiguration
@@ -19228,6 +19244,8 @@
 
 Create the input file for the given directories to 
compact.
 
+createInputFileListing(Job)
 - Method in class org.apache.hadoop.hbase.backup.mapreduce.MapReduceBackupCopyJob.BackupDistCp
+ 
 createIO(MetricsIOWrapper)
 - Method in interface org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactory
 
 Create a MetricsIOSource from a MetricsIOWrapper.
@@ -20939,8 +20957,6 @@
  
 data
 - Variable in class org.apache.hadoop.hbase.client.AsyncProcessTask.ListRowAccess
  
-data
 - Variable in class org.apache.hadoop.hbase.coprocessor.example.ZooKeeperScanPolicyObserver.ZKWatcher
- 
 data
 - Variable in class org.apache.hadoop.hbase.io.hfile.bucket.BucketCache.RAMQueueEntry
  
 data
 - Variable in class org.apache.hadoop.hbase.ipc.SimpleServerRpcConnection
@@ -23253,6 +23269,12 @@
 
 DelegatingHBaseRpcController(HBaseRpcController)
 - Constructor for class org.apache.hadoop.hbase.ipc.DelegatingHBaseRpcController
  
+DelegatingInternalScanner - Class in org.apache.hadoop.hbase.coprocessor.example
+
+A simple delegation for doing filtering on InternalScanner.
+
+DelegatingInternalScanner(InternalScanner)
 - Constructor for class org.apache.hadoop.hbase.coprocessor.example.DelegatingInternalScanner
+ 
 delegator
 - Variable in class org.apache.hadoop.hbase.security.visibility.VisibilityReplicationEndpoint
  
 delete()
 - Method in class org.apache.hadoop.hbase.backup.HFileArchiver.File
@@ -27063,6 +27085,8 @@
 
 Add a task entry to coordination if it is not already 
there.
 
+ensemble
 - Variable in class org.apache.hadoop.