[1/2] hbase git commit: HBASE-18995 Move methods that are for internal usage from CellUtil to Private util class (Ramkrishna Vasudevan); ADDENDUM add file I forgot to add

2017-10-27 Thread stack
Repository: hbase
Updated Branches:
  refs/heads/branch-2 250a8bc7a -> efb5d7b24


http://git-wip-us.apache.org/repos/asf/hbase/blob/efb5d7b2/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java
new file mode 100644
index 000..d70d974
--- /dev/null
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java
@@ -0,0 +1,2776 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.hbase;
+
+import static org.apache.hadoop.hbase.HConstants.EMPTY_BYTE_ARRAY;
+import static org.apache.hadoop.hbase.Tag.TAG_LENGTH_SIZE;
+
+import java.io.DataOutput;
+import java.io.DataOutputStream;
+import java.io.IOException;
+import java.io.OutputStream;
+import java.math.BigDecimal;
+import java.nio.ByteBuffer;
+import java.util.ArrayList;
+import java.util.Iterator;
+import java.util.List;
+
+import org.apache.hadoop.hbase.KeyValue.Type;
+import org.apache.hadoop.hbase.filter.ByteArrayComparable;
+import org.apache.hadoop.hbase.io.HeapSize;
+import org.apache.hadoop.hbase.io.TagCompressionContext;
+import org.apache.hadoop.hbase.io.util.Dictionary;
+import org.apache.hadoop.hbase.io.util.StreamUtils;
+import org.apache.hadoop.hbase.util.ByteBufferUtils;
+import org.apache.hadoop.hbase.util.ByteRange;
+import org.apache.hadoop.hbase.util.Bytes;
+import org.apache.hadoop.hbase.util.ClassSize;
+import org.apache.yetus.audience.InterfaceAudience;
+
+import com.google.common.annotations.VisibleForTesting;
+
+/**
+ * Utility methods helpful slinging {@link Cell} instances. It has more 
powerful and
+ * rich set of APIs than those in {@link CellUtil} for internal usage.
+ */
+@InterfaceAudience.Private
+// TODO : Make Tag IA.LimitedPrivate and move some of the Util methods to CP 
exposed Util class
+public class PrivateCellUtil {
+
+  /**
+   * Private constructor to keep this class from being instantiated.
+   */
+  private PrivateCellUtil() {
+
+  }
+
+  /*** ByteRange ***/
+
+  public static ByteRange fillRowRange(Cell cell, ByteRange range) {
+return range.set(cell.getRowArray(), cell.getRowOffset(), 
cell.getRowLength());
+  }
+
+  public static ByteRange fillFamilyRange(Cell cell, ByteRange range) {
+return range.set(cell.getFamilyArray(), cell.getFamilyOffset(), 
cell.getFamilyLength());
+  }
+
+  public static ByteRange fillQualifierRange(Cell cell, ByteRange range) {
+return range.set(cell.getQualifierArray(), cell.getQualifierOffset(),
+  cell.getQualifierLength());
+  }
+
+  public static ByteRange fillValueRange(Cell cell, ByteRange range) {
+return range.set(cell.getValueArray(), cell.getValueOffset(), 
cell.getValueLength());
+  }
+
+  public static ByteRange fillTagRange(Cell cell, ByteRange range) {
+return range.set(cell.getTagsArray(), cell.getTagsOffset(), 
cell.getTagsLength());
+  }
+
+  /**
+   * Returns tag value in a new byte array. If server-side, use {@link 
Tag#getValueArray()} with
+   * appropriate {@link Tag#getValueOffset()} and {@link Tag#getValueLength()} 
instead to save on
+   * allocations.
+   * @param cell
+   * @return tag value in a new byte array.
+   */
+  public static byte[] getTagsArray(Cell cell) {
+byte[] output = new byte[cell.getTagsLength()];
+copyTagsTo(cell, output, 0);
+return output;
+  }
+
+  public static byte[] cloneTags(Cell cell) {
+byte[] output = new byte[cell.getTagsLength()];
+copyTagsTo(cell, output, 0);
+return output;
+  }
+
+  /**
+   * Copies the tags info into the tag portion of the cell
+   * @param cell
+   * @param destination
+   * @param destinationOffset
+   * @return position after tags
+   */
+  public static int copyTagsTo(Cell cell, byte[] destination, int 
destinationOffset) {
+int tlen = cell.getTagsLength();
+if (cell instanceof ByteBufferCell) {
+  ByteBufferUtils.copyFromBufferToArray(destination,
+((ByteBufferCell) cell).getTagsByteBuffer(), ((ByteBufferCell) 
cell).getTagsPosition(),
+

[2/2] hbase git commit: HBASE-18995 Move methods that are for internal usage from CellUtil to Private util class (Ramkrishna Vasudevan); ADDENDUM add file I forgot to add

2017-10-27 Thread stack
HBASE-18995 Move methods that are for internal usage from CellUtil to Private 
util class (Ramkrishna Vasudevan); ADDENDUM add file I forgot to add


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/efb5d7b2
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/efb5d7b2
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/efb5d7b2

Branch: refs/heads/branch-2
Commit: efb5d7b24df5111cd0f4d4282a1ffd6c98ef405a
Parents: 250a8bc
Author: Michael Stack 
Authored: Fri Oct 27 20:34:59 2017 -0700
Committer: Michael Stack 
Committed: Fri Oct 27 20:35:51 2017 -0700

--
 .../apache/hadoop/hbase/PrivateCellUtil.java| 2776 ++
 1 file changed, 2776 insertions(+)
--




[1/2] hbase git commit: HBASE-18995 Move methods that are for internal usage from CellUtil to Private util class (Ramkrishna Vasudevan); ADDENDUM add file I forgot to add

2017-10-27 Thread stack
Repository: hbase
Updated Branches:
  refs/heads/master f6c2490b8 -> bfaacfdba


http://git-wip-us.apache.org/repos/asf/hbase/blob/bfaacfdb/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java
new file mode 100644
index 000..d70d974
--- /dev/null
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/PrivateCellUtil.java
@@ -0,0 +1,2776 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.hbase;
+
+import static org.apache.hadoop.hbase.HConstants.EMPTY_BYTE_ARRAY;
+import static org.apache.hadoop.hbase.Tag.TAG_LENGTH_SIZE;
+
+import java.io.DataOutput;
+import java.io.DataOutputStream;
+import java.io.IOException;
+import java.io.OutputStream;
+import java.math.BigDecimal;
+import java.nio.ByteBuffer;
+import java.util.ArrayList;
+import java.util.Iterator;
+import java.util.List;
+
+import org.apache.hadoop.hbase.KeyValue.Type;
+import org.apache.hadoop.hbase.filter.ByteArrayComparable;
+import org.apache.hadoop.hbase.io.HeapSize;
+import org.apache.hadoop.hbase.io.TagCompressionContext;
+import org.apache.hadoop.hbase.io.util.Dictionary;
+import org.apache.hadoop.hbase.io.util.StreamUtils;
+import org.apache.hadoop.hbase.util.ByteBufferUtils;
+import org.apache.hadoop.hbase.util.ByteRange;
+import org.apache.hadoop.hbase.util.Bytes;
+import org.apache.hadoop.hbase.util.ClassSize;
+import org.apache.yetus.audience.InterfaceAudience;
+
+import com.google.common.annotations.VisibleForTesting;
+
+/**
+ * Utility methods helpful slinging {@link Cell} instances. It has more 
powerful and
+ * rich set of APIs than those in {@link CellUtil} for internal usage.
+ */
+@InterfaceAudience.Private
+// TODO : Make Tag IA.LimitedPrivate and move some of the Util methods to CP 
exposed Util class
+public class PrivateCellUtil {
+
+  /**
+   * Private constructor to keep this class from being instantiated.
+   */
+  private PrivateCellUtil() {
+
+  }
+
+  /*** ByteRange ***/
+
+  public static ByteRange fillRowRange(Cell cell, ByteRange range) {
+return range.set(cell.getRowArray(), cell.getRowOffset(), 
cell.getRowLength());
+  }
+
+  public static ByteRange fillFamilyRange(Cell cell, ByteRange range) {
+return range.set(cell.getFamilyArray(), cell.getFamilyOffset(), 
cell.getFamilyLength());
+  }
+
+  public static ByteRange fillQualifierRange(Cell cell, ByteRange range) {
+return range.set(cell.getQualifierArray(), cell.getQualifierOffset(),
+  cell.getQualifierLength());
+  }
+
+  public static ByteRange fillValueRange(Cell cell, ByteRange range) {
+return range.set(cell.getValueArray(), cell.getValueOffset(), 
cell.getValueLength());
+  }
+
+  public static ByteRange fillTagRange(Cell cell, ByteRange range) {
+return range.set(cell.getTagsArray(), cell.getTagsOffset(), 
cell.getTagsLength());
+  }
+
+  /**
+   * Returns tag value in a new byte array. If server-side, use {@link 
Tag#getValueArray()} with
+   * appropriate {@link Tag#getValueOffset()} and {@link Tag#getValueLength()} 
instead to save on
+   * allocations.
+   * @param cell
+   * @return tag value in a new byte array.
+   */
+  public static byte[] getTagsArray(Cell cell) {
+byte[] output = new byte[cell.getTagsLength()];
+copyTagsTo(cell, output, 0);
+return output;
+  }
+
+  public static byte[] cloneTags(Cell cell) {
+byte[] output = new byte[cell.getTagsLength()];
+copyTagsTo(cell, output, 0);
+return output;
+  }
+
+  /**
+   * Copies the tags info into the tag portion of the cell
+   * @param cell
+   * @param destination
+   * @param destinationOffset
+   * @return position after tags
+   */
+  public static int copyTagsTo(Cell cell, byte[] destination, int 
destinationOffset) {
+int tlen = cell.getTagsLength();
+if (cell instanceof ByteBufferCell) {
+  ByteBufferUtils.copyFromBufferToArray(destination,
+((ByteBufferCell) cell).getTagsByteBuffer(), ((ByteBufferCell) 
cell).getTagsPosition(),
+  

[2/2] hbase git commit: HBASE-18995 Move methods that are for internal usage from CellUtil to Private util class (Ramkrishna Vasudevan); ADDENDUM add file I forgot to add

2017-10-27 Thread stack
HBASE-18995 Move methods that are for internal usage from CellUtil to Private 
util class (Ramkrishna Vasudevan); ADDENDUM add file I forgot to add


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/bfaacfdb
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/bfaacfdb
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/bfaacfdb

Branch: refs/heads/master
Commit: bfaacfdba306061477d3103754fd5c5ac3612e1b
Parents: f6c2490
Author: Michael Stack 
Authored: Fri Oct 27 20:34:59 2017 -0700
Committer: Michael Stack 
Committed: Fri Oct 27 20:34:59 2017 -0700

--
 .../apache/hadoop/hbase/PrivateCellUtil.java| 2776 ++
 1 file changed, 2776 insertions(+)
--




[6/6] hbase git commit: HBASE-18995 Move methods that are for internal usage from CellUtil to Private util class (Ramkrishna Vasudevan)

2017-10-27 Thread stack
HBASE-18995 Move methods that are for internal usage from CellUtil to Private 
util class (Ramkrishna Vasudevan)


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/250a8bc7
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/250a8bc7
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/250a8bc7

Branch: refs/heads/branch-2
Commit: 250a8bc7ad05080692b461a28f6fe1fa614b5830
Parents: f981de5
Author: Michael Stack 
Authored: Fri Oct 27 17:51:32 2017 -0700
Committer: Michael Stack 
Committed: Fri Oct 27 17:51:32 2017 -0700

--
 .../client/AllowPartialScanResultCache.java |2 +-
 .../hbase/client/BatchScanResultCache.java  |2 +-
 .../hadoop/hbase/client/ConnectionUtils.java|8 +-
 .../org/apache/hadoop/hbase/client/Delete.java  |3 +-
 .../apache/hadoop/hbase/client/Increment.java   |2 +-
 .../apache/hadoop/hbase/client/Mutation.java|5 +-
 .../org/apache/hadoop/hbase/client/Put.java |2 +-
 .../org/apache/hadoop/hbase/client/Result.java  |8 +-
 .../BigDecimalColumnInterpreter.java|3 +-
 .../coprocessor/DoubleColumnInterpreter.java|4 +-
 .../coprocessor/LongColumnInterpreter.java  |4 +-
 .../hbase/filter/ColumnPaginationFilter.java|3 +-
 .../hadoop/hbase/filter/ColumnPrefixFilter.java |4 +-
 .../hadoop/hbase/filter/ColumnRangeFilter.java  |3 +-
 .../hadoop/hbase/filter/CompareFilter.java  |   18 +-
 .../hadoop/hbase/filter/FuzzyRowFilter.java |3 +-
 .../hbase/filter/MultiRowRangeFilter.java   |3 +-
 .../filter/MultipleColumnPrefixFilter.java  |3 +-
 .../hbase/filter/SingleColumnValueFilter.java   |3 +-
 .../hadoop/hbase/filter/TimestampsFilter.java   |6 +-
 .../hadoop/hbase/filter/TestComparators.java|   37 +-
 .../hadoop/hbase/ipc/TestCellBlockBuilder.java  |3 +-
 .../java/org/apache/hadoop/hbase/CellUtil.java  | 4196 +-
 .../hadoop/hbase/IndividualBytesFieldCell.java  |3 +-
 .../java/org/apache/hadoop/hbase/KeyValue.java  |4 +-
 .../apache/hadoop/hbase/KeyValueTestUtil.java   |2 +-
 .../org/apache/hadoop/hbase/KeyValueUtil.java   |4 +-
 .../java/org/apache/hadoop/hbase/TagUtil.java   |2 +-
 .../io/encoding/BufferedDataBlockEncoder.java   |7 +-
 .../hbase/io/encoding/DiffKeyDeltaEncoder.java  |   14 +-
 .../hbase/io/encoding/FastDiffDeltaEncoder.java |   18 +-
 .../hadoop/hbase/io/encoding/NoneEncoder.java   |4 +-
 .../io/encoding/PrefixKeyDeltaEncoder.java  |   16 +-
 .../hbase/io/encoding/RowIndexSeekerV1.java |5 +-
 .../hadoop/hbase/util/RowBloomHashKey.java  |3 +-
 .../hadoop/hbase/util/RowColBloomHashKey.java   |5 +-
 .../hadoop/hbase/TestByteBufferKeyValue.java|4 +-
 .../apache/hadoop/hbase/TestCellComparator.java |   15 +-
 .../org/apache/hadoop/hbase/TestCellUtil.java   |   97 +-
 .../org/apache/hadoop/hbase/TestKeyValue.java   |4 +-
 .../hbase/mapreduce/CellSerialization.java  |7 +-
 .../hadoop/hbase/mapreduce/CellSortReducer.java |4 +-
 .../hbase/mapreduce/HFileOutputFormat2.java |6 +-
 .../apache/hadoop/hbase/mapreduce/Import.java   |   17 +-
 .../mapreduce/TableSnapshotInputFormatImpl.java |8 +-
 .../hadoop/hbase/mapreduce/WALPlayer.java   |2 +-
 .../apache/hadoop/hbase/util/MapReduceCell.java |   17 +-
 .../TestCellBasedHFileOutputFormat2.java|2 +-
 .../mapreduce/TestCellBasedImportExport2.java   |3 +-
 .../hbase/mapreduce/TestHFileOutputFormat2.java |2 +-
 .../hbase/mapreduce/TestImportExport.java   |3 +-
 .../TestImportTSVWithOperationAttributes.java   |4 +-
 .../TestImportTSVWithVisibilityLabels.java  |4 +-
 .../hadoop/hbase/mapreduce/TestImportTsv.java   |4 +-
 .../hadoop/hbase/mapreduce/TestSyncTable.java   |2 +-
 .../codec/prefixtree/PrefixTreeSeeker.java  |3 +-
 .../decode/PrefixTreeArrayScanner.java  |3 +-
 .../decode/PrefixTreeArraySearcher.java |7 +-
 .../codec/prefixtree/decode/PrefixTreeCell.java |4 +-
 .../prefixtree/encode/PrefixTreeEncoder.java|9 +-
 .../codec/prefixtree/row/BaseTestRowData.java   |2 +-
 .../prefixtree/row/TestPrefixTreeSearcher.java  |3 +-
 .../hbase/client/ClientSideRegionScanner.java   |3 +-
 .../hbase/client/TableSnapshotScanner.java  |3 +-
 .../hadoop/hbase/io/HalfStoreFileReader.java|   13 +-
 .../io/hfile/CompoundBloomFilterWriter.java |4 +-
 .../hadoop/hbase/io/hfile/HFileBlockIndex.java  |5 +-
 .../hbase/io/hfile/HFilePrettyPrinter.java  |3 +-
 .../hadoop/hbase/io/hfile/HFileReaderImpl.java  |   10 +-
 .../hadoop/hbase/io/hfile/HFileWriterImpl.java  |   15 +-
 .../hbase/mob/DefaultMobStoreCompactor.java |3 +-
 .../org/apache/hadoop/hbase/mob/MobUtils.java   |   11 +-
 .../hbase/protobuf/ReplicationProtbufUtil.java  |4 +-
 ...

[4/6] hbase git commit: HBASE-18995 Move methods that are for internal usage from CellUtil to Private util class (Ramkrishna Vasudevan)

2017-10-27 Thread stack
http://git-wip-us.apache.org/repos/asf/hbase/blob/250a8bc7/hbase-common/src/main/java/org/apache/hadoop/hbase/IndividualBytesFieldCell.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/IndividualBytesFieldCell.java
 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/IndividualBytesFieldCell.java
index 14e35df..a15843c 100644
--- 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/IndividualBytesFieldCell.java
+++ 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/IndividualBytesFieldCell.java
@@ -24,7 +24,6 @@ import java.io.IOException;
 import java.io.OutputStream;
 import java.nio.ByteBuffer;
 
-import org.apache.yetus.audience.InterfaceAudience;
 import org.apache.hadoop.hbase.util.ArrayUtils;
 import org.apache.hadoop.hbase.util.ByteBufferUtils;
 import org.apache.hadoop.hbase.util.Bytes;
@@ -144,7 +143,7 @@ public class IndividualBytesFieldCell implements 
ExtendedCell {
 ByteBufferUtils.putInt(out, getValueLength());
 
 // Key
-CellUtil.writeFlatKey(this, out);
+PrivateCellUtil.writeFlatKey(this, out);
 
 // Value
 out.write(getValueArray());

http://git-wip-us.apache.org/repos/asf/hbase/blob/250a8bc7/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
--
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
index 6154045..42ac97d 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
@@ -1617,7 +1617,7 @@ public class KeyValue implements ExtendedCell {
  */
 @Override
 public int compare(final Cell left, final Cell right) {
-  return 
CellUtil.compareKeyIgnoresMvcc(CellComparatorImpl.META_COMPARATOR, left, right);
+  return 
PrivateCellUtil.compareKeyIgnoresMvcc(CellComparatorImpl.META_COMPARATOR, left, 
right);
 }
 
 @Override
@@ -1839,7 +1839,7 @@ public class KeyValue implements ExtendedCell {
 }
 
 public int compareOnlyKeyPortion(Cell left, Cell right) {
-  return CellUtil.compareKeyIgnoresMvcc(CellComparatorImpl.COMPARATOR, 
left, right);
+  return 
PrivateCellUtil.compareKeyIgnoresMvcc(CellComparatorImpl.COMPARATOR, left, 
right);
 }
 
 /**

http://git-wip-us.apache.org/repos/asf/hbase/blob/250a8bc7/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueTestUtil.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueTestUtil.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueTestUtil.java
index c8bed3e..7467d67 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueTestUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueTestUtil.java
@@ -85,7 +85,7 @@ public class KeyValueTestUtil {
 for (Cell kv1 : kvCollection1) {
   boolean found = false;
   for (Cell kv2 : kvCollection2) {
-if (CellUtil.equalsIgnoreMvccVersion(kv1, kv2)) found = true;
+if (PrivateCellUtil.equalsIgnoreMvccVersion(kv1, kv2)) found = true;
   }
   if (!found) return false;
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/250a8bc7/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java
index f1f03eb..6fd37c0 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java
@@ -163,7 +163,7 @@ public class KeyValueUtil {
 pos = CellUtil.copyValueTo(cell, output, pos);
 if (withTags && (cell.getTagsLength() > 0)) {
   pos = Bytes.putAsShort(output, pos, cell.getTagsLength());
-  pos = CellUtil.copyTagTo(cell, output, pos);
+  pos = PrivateCellUtil.copyTagsTo(cell, output, pos);
 }
 return pos;
   }
@@ -179,7 +179,7 @@ public class KeyValueUtil {
 int tagsLength = cell.getTagsLength();
 if (withTags && (tagsLength > 0)) {
   offset = ByteBufferUtils.putAsShort(buf, offset, tagsLength);// Tags 
length
-  offset = CellUtil.copyTagTo(cell, buf, offset);// Tags bytes
+  offset = PrivateCellUtil.copyTagsTo(cell, buf, offset);// Tags bytes
 }
 return offset;
   }

http://git-wip-us.apache.org/repos/asf/hbase/blob/250a8bc7/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java
--
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java
index 0f010a0..a4962f4 100644
-

[5/6] hbase git commit: HBASE-18995 Move methods that are for internal usage from CellUtil to Private util class (Ramkrishna Vasudevan)

2017-10-27 Thread stack
http://git-wip-us.apache.org/repos/asf/hbase/blob/250a8bc7/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java
--
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java
index f9640a3..c2fb869 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java
@@ -18,7 +18,6 @@
 
 package org.apache.hadoop.hbase;
 
-import static org.apache.hadoop.hbase.HConstants.EMPTY_BYTE_ARRAY;
 import static org.apache.hadoop.hbase.Tag.TAG_LENGTH_SIZE;
 import static org.apache.hadoop.hbase.KeyValue.COLUMN_FAMILY_DELIMITER;
 import static org.apache.hadoop.hbase.KeyValue.getDelimiter;
@@ -27,8 +26,6 @@ import static 
org.apache.hadoop.hbase.KeyValue.COLUMN_FAMILY_DELIM_ARRAY;
 import java.io.DataOutput;
 import java.io.DataOutputStream;
 import java.io.IOException;
-import java.io.OutputStream;
-import java.math.BigDecimal;
 import java.nio.ByteBuffer;
 import java.util.ArrayList;
 import java.util.Arrays;
@@ -38,25 +35,20 @@ import java.util.Map.Entry;
 import java.util.NavigableMap;
 
 import org.apache.hadoop.hbase.KeyValue.Type;
-import org.apache.hadoop.hbase.filter.ByteArrayComparable;
 import org.apache.yetus.audience.InterfaceAudience;
 import org.apache.yetus.audience.InterfaceAudience.Private;
 
 import com.google.common.annotations.VisibleForTesting;
 
 import org.apache.hadoop.hbase.io.HeapSize;
-import org.apache.hadoop.hbase.io.TagCompressionContext;
-import org.apache.hadoop.hbase.io.util.Dictionary;
-import org.apache.hadoop.hbase.io.util.StreamUtils;
 import org.apache.hadoop.hbase.util.ByteBufferUtils;
 import org.apache.hadoop.hbase.util.ByteRange;
 import org.apache.hadoop.hbase.util.Bytes;
-import org.apache.hadoop.hbase.util.ClassSize;
 
 /**
- * Utility methods helpful slinging {@link Cell} instances.
- * Some methods below are for internal use only and are marked 
InterfaceAudience.Private at the
- * method level.
+ * Utility methods helpful for slinging {@link Cell} instances. Some methods 
below are for internal
+ * use only and are marked InterfaceAudience.Private at the method level. Note 
that all such methods
+ * have been marked deprecated in HBase-2.0 which will be subsequently removed 
in HBase-3.0
  */
 @InterfaceAudience.Public
 public final class CellUtil {
@@ -64,73 +56,100 @@ public final class CellUtil {
   /**
* Private constructor to keep this class from being instantiated.
*/
-  private CellUtil(){}
+  private CellUtil() {
+  }
 
   /*** ByteRange ***/
 
+  /**
+   * @deprecated As of HBase-2.0. Will be removed in HBase-3.0.
+   */
+  @Deprecated
   public static ByteRange fillRowRange(Cell cell, ByteRange range) {
 return range.set(cell.getRowArray(), cell.getRowOffset(), 
cell.getRowLength());
   }
 
+  /**
+   * @deprecated As of HBase-2.0. Will be removed in HBase-3.0.
+   */
+  @Deprecated
   public static ByteRange fillFamilyRange(Cell cell, ByteRange range) {
 return range.set(cell.getFamilyArray(), cell.getFamilyOffset(), 
cell.getFamilyLength());
   }
 
+  /**
+   * @deprecated As of HBase-2.0. Will be removed in HBase-3.0.
+   */
+  @Deprecated
   public static ByteRange fillQualifierRange(Cell cell, ByteRange range) {
 return range.set(cell.getQualifierArray(), cell.getQualifierOffset(),
   cell.getQualifierLength());
   }
 
+  /**
+   * @deprecated As of HBase-2.0. Will be removed in HBase-3.0.
+   */
+  @Deprecated
   public static ByteRange fillValueRange(Cell cell, ByteRange range) {
 return range.set(cell.getValueArray(), cell.getValueOffset(), 
cell.getValueLength());
   }
 
+  /**
+   * @deprecated As of HBase-2.0. Will be removed in HBase-3.0.
+   */
+  @Deprecated
   public static ByteRange fillTagRange(Cell cell, ByteRange range) {
 return range.set(cell.getTagsArray(), cell.getTagsOffset(), 
cell.getTagsLength());
   }
 
   /* get individual arrays for tests /
 
-  public static byte[] cloneRow(Cell cell){
+  public static byte[] cloneRow(Cell cell) {
 byte[] output = new byte[cell.getRowLength()];
 copyRowTo(cell, output, 0);
 return output;
   }
 
-  public static byte[] cloneFamily(Cell cell){
+  public static byte[] cloneFamily(Cell cell) {
 byte[] output = new byte[cell.getFamilyLength()];
 copyFamilyTo(cell, output, 0);
 return output;
   }
 
-  public static byte[] cloneQualifier(Cell cell){
+  public static byte[] cloneQualifier(Cell cell) {
 byte[] output = new byte[cell.getQualifierLength()];
 copyQualifierTo(cell, output, 0);
 return output;
   }
 
-  public static byte[] cloneValue(Cell cell){
+  public static byte[] cloneValue(Cell cell) {
 byte[] output = new byte[cell.getValueLength()];
 copyValueTo(cell, output, 0);
 return output;
   }
 
+  /

[3/6] hbase git commit: HBASE-18995 Move methods that are for internal usage from CellUtil to Private util class (Ramkrishna Vasudevan)

2017-10-27 Thread stack
http://git-wip-us.apache.org/repos/asf/hbase/blob/250a8bc7/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java
index 7068fe1..f2416bc 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java
@@ -37,6 +37,7 @@ import org.apache.hadoop.hbase.CellComparator;
 import org.apache.hadoop.hbase.CellComparatorImpl;
 import org.apache.hadoop.hbase.CellUtil;
 import org.apache.hadoop.hbase.HConstants;
+import org.apache.hadoop.hbase.PrivateCellUtil;
 import org.apache.hadoop.hbase.KeyValue;
 import org.apache.hadoop.hbase.ByteBufferKeyValue;
 import org.apache.hadoop.hbase.SizeCachedKeyValue;
@@ -728,7 +729,8 @@ public class HFileReaderImpl implements HFile.Reader, 
Configurable {
 offsetFromPos += Bytes.SIZEOF_LONG;
 blockBuffer.asSubByteBuffer(blockBuffer.position() + offsetFromPos, 
klen, pair);
 bufBackedKeyOnlyKv.setKey(pair.getFirst(), pair.getSecond(), klen);
-int comp = CellUtil.compareKeyIgnoresMvcc(reader.getComparator(), key, 
bufBackedKeyOnlyKv);
+int comp =
+PrivateCellUtil.compareKeyIgnoresMvcc(reader.getComparator(), key, 
bufBackedKeyOnlyKv);
 offsetFromPos += klen + vlen;
 if (this.reader.getFileContext().isIncludesTags()) {
   // Read short as unsigned, high byte first
@@ -811,7 +813,7 @@ public class HFileReaderImpl implements HFile.Reader, 
Configurable {
 } else {
   // The comparison with no_next_index_key has to be checked
   if (this.nextIndexedKey != null &&
-  (this.nextIndexedKey == KeyValueScanner.NO_NEXT_INDEXED_KEY || 
CellUtil
+  (this.nextIndexedKey == KeyValueScanner.NO_NEXT_INDEXED_KEY || 
PrivateCellUtil
   .compareKeyIgnoresMvcc(reader.getComparator(), key, 
nextIndexedKey) < 0)) {
 // The reader shall continue to scan the current data block instead
 // of querying the
@@ -865,7 +867,7 @@ public class HFileReaderImpl implements HFile.Reader, 
Configurable {
 return false;
   }
   Cell firstKey = getFirstKeyCellInBlock(seekToBlock);
-  if (CellUtil.compareKeyIgnoresMvcc(reader.getComparator(), firstKey, 
key) >= 0) {
+  if (PrivateCellUtil.compareKeyIgnoresMvcc(reader.getComparator(), 
firstKey, key) >= 0) {
 long previousBlockOffset = seekToBlock.getPrevBlockOffset();
 // The key we are interested in
 if (previousBlockOffset == -1) {
@@ -1229,7 +1231,7 @@ public class HFileReaderImpl implements HFile.Reader, 
Configurable {
 public int compareKey(CellComparator comparator, Cell key) {
   blockBuffer.asSubByteBuffer(blockBuffer.position() + KEY_VALUE_LEN_SIZE, 
currKeyLen, pair);
   this.bufBackedKeyOnlyKv.setKey(pair.getFirst(), pair.getSecond(), 
currKeyLen);
-  return CellUtil.compareKeyIgnoresMvcc(comparator, key, 
this.bufBackedKeyOnlyKv);
+  return PrivateCellUtil.compareKeyIgnoresMvcc(comparator, key, 
this.bufBackedKeyOnlyKv);
 }
 
 @Override

http://git-wip-us.apache.org/repos/asf/hbase/blob/250a8bc7/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterImpl.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterImpl.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterImpl.java
index 33cfa1d..bd98cdd 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterImpl.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterImpl.java
@@ -39,6 +39,7 @@ import org.apache.hadoop.hbase.CellComparator;
 import org.apache.hadoop.hbase.CellComparatorImpl;
 import org.apache.hadoop.hbase.CellUtil;
 import org.apache.hadoop.hbase.HConstants;
+import org.apache.hadoop.hbase.PrivateCellUtil;
 import org.apache.hadoop.hbase.KeyValueUtil;
 import org.apache.hadoop.hbase.CellComparatorImpl.MetaCellComparator;
 import org.apache.yetus.audience.InterfaceAudience;
@@ -239,7 +240,7 @@ public class HFileWriterImpl implements HFile.Writer {
   throw new IOException("Key cannot be null or empty");
 }
 if (lastCell != null) {
-  int keyComp = CellUtil.compareKeyIgnoresMvcc(comparator, lastCell, cell);
+  int keyComp = PrivateCellUtil.compareKeyIgnoresMvcc(comparator, 
lastCell, cell);
 
   if (keyComp > 0) {
 throw new IOException("Added a key not lexically larger than"
@@ -341,7 +342,7 @@ public class HFileWriterImpl implements HFile.Writer {
 int onDiskSize = blockWriter.getOnDiskSizeWithHeader();
 Cell indexEntry =
   get

[1/6] hbase git commit: HBASE-18995 Move methods that are for internal usage from CellUtil to Private util class (Ramkrishna Vasudevan)

2017-10-27 Thread stack
Repository: hbase
Updated Branches:
  refs/heads/branch-2 f981de5bb -> 250a8bc7a


http://git-wip-us.apache.org/repos/asf/hbase/blob/250a8bc7/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java
--
diff --git 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java
 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java
index 6224495..e0f2986 100644
--- 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java
+++ 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java
@@ -262,7 +262,7 @@ public class TestCoprocessorScanPolicy {
   if (version == 0) {
 return c -> true;
   } else {
-if (row == null || !CellUtil.matchingRow(firstCell, row)) {
+if (row == null || !CellUtil.matchingRows(firstCell, row)) {
   row = CellUtil.cloneRow(firstCell);
   // reset qualifier as there is a row change
   qualifier = null;

http://git-wip-us.apache.org/repos/asf/hbase/blob/250a8bc7/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftUtilities.java
--
diff --git 
a/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftUtilities.java
 
b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftUtilities.java
index 271b70c..4e15a9c 100644
--- 
a/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftUtilities.java
+++ 
b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftUtilities.java
@@ -30,6 +30,7 @@ import org.apache.commons.collections4.MapUtils;
 import org.apache.hadoop.hbase.CompareOperator;
 import org.apache.hadoop.hbase.HRegionInfo;
 import org.apache.hadoop.hbase.HRegionLocation;
+import org.apache.hadoop.hbase.PrivateCellUtil;
 import org.apache.hadoop.hbase.ServerName;
 import org.apache.yetus.audience.InterfaceAudience;
 import org.apache.hadoop.hbase.Cell;
@@ -171,7 +172,7 @@ public class ThriftUtilities {
   col.setTimestamp(kv.getTimestamp());
   col.setValue(CellUtil.cloneValue(kv));
   if (kv.getTagsLength() > 0) {
-col.setTags(CellUtil.getTagArray(kv));
+col.setTags(PrivateCellUtil.getTagsArray(kv));
   }
   columnValues.add(col);
 }



[2/6] hbase git commit: HBASE-18995 Move methods that are for internal usage from CellUtil to Private util class (Ramkrishna Vasudevan)

2017-10-27 Thread stack
http://git-wip-us.apache.org/repos/asf/hbase/blob/250a8bc7/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java
index a3841aa..84edf37 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java
@@ -44,13 +44,13 @@ import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.hbase.AuthUtil;
 import org.apache.hadoop.hbase.Cell;
 import org.apache.hadoop.hbase.CellScanner;
-import org.apache.hadoop.hbase.CellUtil;
 import org.apache.hadoop.hbase.CoprocessorEnvironment;
 import org.apache.hadoop.hbase.DoNotRetryIOException;
 import org.apache.hadoop.hbase.HBaseInterfaceAudience;
 import org.apache.hadoop.hbase.HColumnDescriptor;
 import org.apache.hadoop.hbase.HConstants;
 import org.apache.hadoop.hbase.HTableDescriptor;
+import org.apache.hadoop.hbase.PrivateCellUtil;
 import org.apache.hadoop.hbase.MetaTableAccessor;
 import org.apache.hadoop.hbase.TableName;
 import org.apache.hadoop.hbase.Tag;
@@ -58,7 +58,6 @@ import org.apache.hadoop.hbase.TagType;
 import org.apache.hadoop.hbase.TagUtil;
 import org.apache.hadoop.hbase.client.Admin;
 import org.apache.hadoop.hbase.client.Append;
-import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
 import org.apache.hadoop.hbase.client.Delete;
 import org.apache.hadoop.hbase.client.Get;
 import org.apache.hadoop.hbase.client.Increment;
@@ -369,13 +368,13 @@ public class VisibilityController implements 
MasterCoprocessor, RegionCoprocesso
 List updatedCells = new ArrayList<>();
 for (CellScanner cellScanner = m.cellScanner(); 
cellScanner.advance();) {
   Cell cell = cellScanner.current();
-  List tags = CellUtil.getTags(cell);
+  List tags = PrivateCellUtil.getTags(cell);
   if (modifiedTagFound) {
 // Rewrite the tags by removing the modified tags.
 removeReplicationVisibilityTag(tags);
   }
   tags.addAll(visibilityTags);
-  Cell updatedCell = CellUtil.createCell(cell, tags);
+  Cell updatedCell = PrivateCellUtil.createCell(cell, tags);
   updatedCells.add(updatedCell);
 }
 m.getFamilyCellMap().clear();
@@ -428,7 +427,7 @@ public class VisibilityController implements 
MasterCoprocessor, RegionCoprocesso
 
 if (result.size() < get.getMaxVersions()) {
   // Nothing to delete
-  CellUtil.updateLatestStamp(cell, byteNow, 0);
+  PrivateCellUtil.updateLatestStamp(cell, byteNow, 0);
   return;
 }
 if (result.size() > get.getMaxVersions()) {
@@ -436,7 +435,7 @@ public class VisibilityController implements 
MasterCoprocessor, RegionCoprocesso
   + ". Results more than the max versions obtained.");
 }
 Cell getCell = result.get(get.getMaxVersions() - 1);
-CellUtil.setTimestamp(cell, getCell.getTimestamp());
+PrivateCellUtil.setTimestamp(cell, getCell.getTimestamp());
 
 // We are bypassing here because in the 
HRegion.updateDeleteLatestVersionTimeStamp we would
 // update with the current timestamp after again doing a get. As the hook 
as already determined
@@ -472,7 +471,7 @@ public class VisibilityController implements 
MasterCoprocessor, RegionCoprocesso
   // cell visiblilty tags
   // have been modified
   Tag modifiedTag = null;
-  Iterator tagsIterator = CellUtil.tagsIterator(cell);
+  Iterator tagsIterator = PrivateCellUtil.tagsIterator(cell);
   while (tagsIterator.hasNext()) {
 Tag tag = tagsIterator.next();
 if (tag.getType() == TagType.STRING_VIS_TAG_TYPE) {
@@ -484,7 +483,7 @@ public class VisibilityController implements 
MasterCoprocessor, RegionCoprocesso
   pair.setSecond(modifiedTag);
   return pair;
 }
-Iterator tagsItr = CellUtil.tagsIterator(cell);
+Iterator tagsItr = PrivateCellUtil.tagsIterator(cell);
 while (tagsItr.hasNext()) {
   if (RESERVED_VIS_TAG_TYPES.contains(tagsItr.next().getType())) {
 return pair;
@@ -514,7 +513,7 @@ public class VisibilityController implements 
MasterCoprocessor, RegionCoprocesso
 if (isSystemOrSuperUser()) {
   return true;
 }
-Iterator tagsItr = CellUtil.tagsIterator(cell);
+Iterator tagsItr = PrivateCellUtil.tagsIterator(cell);
 while (tagsItr.hasNext()) {
   if (RESERVED_VIS_TAG_TYPES.contains(tagsItr.next().getType())) {
 return false;
@@ -731,7 +730,7 @@ public class VisibilityController implements 
MasterCoprocessor, RegionCoprocesso
 
tags.addAll(this.visibilityLabel

[5/6] hbase git commit: HBASE-18995 Move methods that are for internal usage from CellUtil to Private util class (Ramkrishna Vasudevan)

2017-10-27 Thread stack
http://git-wip-us.apache.org/repos/asf/hbase/blob/f6c2490b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java
--
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java
index 12f6a30..20c217f 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellUtil.java
@@ -18,7 +18,6 @@
 
 package org.apache.hadoop.hbase;
 
-import static org.apache.hadoop.hbase.HConstants.EMPTY_BYTE_ARRAY;
 import static org.apache.hadoop.hbase.Tag.TAG_LENGTH_SIZE;
 import static org.apache.hadoop.hbase.KeyValue.COLUMN_FAMILY_DELIMITER;
 import static org.apache.hadoop.hbase.KeyValue.getDelimiter;
@@ -28,9 +27,7 @@ import java.io.DataOutput;
 import java.io.DataOutputStream;
 import java.io.IOException;
 import java.io.OutputStream;
-import java.math.BigDecimal;
 import java.nio.ByteBuffer;
-import java.util.ArrayList;
 import java.util.Arrays;
 import java.util.Iterator;
 import java.util.List;
@@ -38,25 +35,20 @@ import java.util.Map.Entry;
 import java.util.NavigableMap;
 
 import org.apache.hadoop.hbase.KeyValue.Type;
-import org.apache.hadoop.hbase.filter.ByteArrayComparable;
 import org.apache.yetus.audience.InterfaceAudience;
 import org.apache.yetus.audience.InterfaceAudience.Private;
 
 import com.google.common.annotations.VisibleForTesting;
 
 import org.apache.hadoop.hbase.io.HeapSize;
-import org.apache.hadoop.hbase.io.TagCompressionContext;
-import org.apache.hadoop.hbase.io.util.Dictionary;
-import org.apache.hadoop.hbase.io.util.StreamUtils;
 import org.apache.hadoop.hbase.util.ByteBufferUtils;
 import org.apache.hadoop.hbase.util.ByteRange;
 import org.apache.hadoop.hbase.util.Bytes;
-import org.apache.hadoop.hbase.util.ClassSize;
 
 /**
- * Utility methods helpful slinging {@link Cell} instances.
- * Some methods below are for internal use only and are marked 
InterfaceAudience.Private at the
- * method level.
+ * Utility methods helpful for slinging {@link Cell} instances. Some methods 
below are for internal
+ * use only and are marked InterfaceAudience.Private at the method level. Note 
that all such methods
+ * have been marked deprecated in HBase-2.0 which will be subsequently removed 
in HBase-3.0
  */
 @InterfaceAudience.Public
 public final class CellUtil {
@@ -64,74 +56,96 @@ public final class CellUtil {
   /**
* Private constructor to keep this class from being instantiated.
*/
-  private CellUtil(){}
+  private CellUtil() {
+  }
 
   /*** ByteRange ***/
 
+  /**
+   * @deprecated As of HBase-2.0. Will be removed in HBase-3.0.
+   */
+  @Deprecated
   public static ByteRange fillRowRange(Cell cell, ByteRange range) {
-return range.set(cell.getRowArray(), cell.getRowOffset(), 
cell.getRowLength());
+return PrivateCellUtil.fillRowRange(cell, range);
   }
 
+  /**
+   * @deprecated As of HBase-2.0. Will be removed in HBase-3.0.
+   */
+  @Deprecated
   public static ByteRange fillFamilyRange(Cell cell, ByteRange range) {
-return range.set(cell.getFamilyArray(), cell.getFamilyOffset(), 
cell.getFamilyLength());
+return PrivateCellUtil.fillFamilyRange(cell, range);
   }
 
+  /**
+   * @deprecated As of HBase-2.0. Will be removed in HBase-3.0.
+   */
+  @Deprecated
   public static ByteRange fillQualifierRange(Cell cell, ByteRange range) {
-return range.set(cell.getQualifierArray(), cell.getQualifierOffset(),
-  cell.getQualifierLength());
+return PrivateCellUtil.fillQualifierRange(cell, range);
   }
 
+  /**
+   * @deprecated As of HBase-2.0. Will be removed in HBase-3.0.
+   */
+  @Deprecated
   public static ByteRange fillValueRange(Cell cell, ByteRange range) {
-return range.set(cell.getValueArray(), cell.getValueOffset(), 
cell.getValueLength());
+return PrivateCellUtil.fillValueRange(cell, range);
   }
 
+  /**
+   * @deprecated As of HBase-2.0. Will be removed in HBase-3.0.
+   */
+  @Deprecated
   public static ByteRange fillTagRange(Cell cell, ByteRange range) {
-return range.set(cell.getTagsArray(), cell.getTagsOffset(), 
cell.getTagsLength());
+return PrivateCellUtil.fillTagRange(cell, range);
   }
 
   /* get individual arrays for tests /
 
-  public static byte[] cloneRow(Cell cell){
+  public static byte[] cloneRow(Cell cell) {
 byte[] output = new byte[cell.getRowLength()];
 copyRowTo(cell, output, 0);
 return output;
   }
 
-  public static byte[] cloneFamily(Cell cell){
+  public static byte[] cloneFamily(Cell cell) {
 byte[] output = new byte[cell.getFamilyLength()];
 copyFamilyTo(cell, output, 0);
 return output;
   }
 
-  public static byte[] cloneQualifier(Cell cell){
+  public static byte[] cloneQualifier(Cell cell) {
 byte[] output = new byte[cell.getQualifierLength()];
 copyQualifierTo

[6/6] hbase git commit: HBASE-18995 Move methods that are for internal usage from CellUtil to Private util class (Ramkrishna Vasudevan)

2017-10-27 Thread stack
HBASE-18995 Move methods that are for internal usage from CellUtil to Private 
util class (Ramkrishna Vasudevan)


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/f6c2490b
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/f6c2490b
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/f6c2490b

Branch: refs/heads/master
Commit: f6c2490b82968f33e7d7db0b96896eebdf745e19
Parents: 4dee4a8
Author: Michael Stack 
Authored: Fri Oct 27 17:48:55 2017 -0700
Committer: Michael Stack 
Committed: Fri Oct 27 17:48:55 2017 -0700

--
 .../client/AllowPartialScanResultCache.java |2 +-
 .../hbase/client/BatchScanResultCache.java  |2 +-
 .../hadoop/hbase/client/ConnectionUtils.java|8 +-
 .../org/apache/hadoop/hbase/client/Delete.java  |3 +-
 .../apache/hadoop/hbase/client/Increment.java   |2 +-
 .../apache/hadoop/hbase/client/Mutation.java|5 +-
 .../org/apache/hadoop/hbase/client/Put.java |2 +-
 .../org/apache/hadoop/hbase/client/Result.java  |8 +-
 .../BigDecimalColumnInterpreter.java|3 +-
 .../coprocessor/DoubleColumnInterpreter.java|4 +-
 .../coprocessor/LongColumnInterpreter.java  |4 +-
 .../hbase/filter/ColumnPaginationFilter.java|3 +-
 .../hadoop/hbase/filter/ColumnPrefixFilter.java |4 +-
 .../hadoop/hbase/filter/ColumnRangeFilter.java  |3 +-
 .../hadoop/hbase/filter/CompareFilter.java  |   18 +-
 .../hadoop/hbase/filter/FuzzyRowFilter.java |3 +-
 .../hbase/filter/MultiRowRangeFilter.java   |3 +-
 .../filter/MultipleColumnPrefixFilter.java  |3 +-
 .../hbase/filter/SingleColumnValueFilter.java   |3 +-
 .../hadoop/hbase/filter/TimestampsFilter.java   |6 +-
 .../hadoop/hbase/filter/TestComparators.java|   37 +-
 .../hadoop/hbase/ipc/TestCellBlockBuilder.java  |3 +-
 .../java/org/apache/hadoop/hbase/CellUtil.java  | 3987 --
 .../hadoop/hbase/IndividualBytesFieldCell.java  |3 +-
 .../java/org/apache/hadoop/hbase/KeyValue.java  |4 +-
 .../apache/hadoop/hbase/KeyValueTestUtil.java   |2 +-
 .../org/apache/hadoop/hbase/KeyValueUtil.java   |4 +-
 .../java/org/apache/hadoop/hbase/TagUtil.java   |2 +-
 .../io/encoding/BufferedDataBlockEncoder.java   |7 +-
 .../hbase/io/encoding/DiffKeyDeltaEncoder.java  |   14 +-
 .../hbase/io/encoding/FastDiffDeltaEncoder.java |   18 +-
 .../hadoop/hbase/io/encoding/NoneEncoder.java   |4 +-
 .../io/encoding/PrefixKeyDeltaEncoder.java  |   16 +-
 .../hbase/io/encoding/RowIndexSeekerV1.java |5 +-
 .../hadoop/hbase/util/RowBloomHashKey.java  |3 +-
 .../hadoop/hbase/util/RowColBloomHashKey.java   |5 +-
 .../hadoop/hbase/TestByteBufferKeyValue.java|4 +-
 .../apache/hadoop/hbase/TestCellComparator.java |   15 +-
 .../org/apache/hadoop/hbase/TestCellUtil.java   |   97 +-
 .../org/apache/hadoop/hbase/TestKeyValue.java   |4 +-
 .../hbase/mapreduce/CellSerialization.java  |7 +-
 .../hadoop/hbase/mapreduce/CellSortReducer.java |4 +-
 .../hbase/mapreduce/HFileOutputFormat2.java |6 +-
 .../apache/hadoop/hbase/mapreduce/Import.java   |   16 +-
 .../mapreduce/TableSnapshotInputFormatImpl.java |8 +-
 .../hadoop/hbase/mapreduce/WALPlayer.java   |2 +-
 .../apache/hadoop/hbase/util/MapReduceCell.java |   17 +-
 .../hbase/mapreduce/TestHFileOutputFormat2.java |2 +-
 .../hbase/mapreduce/TestImportExport.java   |3 +-
 .../TestImportTSVWithOperationAttributes.java   |4 +-
 .../TestImportTSVWithVisibilityLabels.java  |4 +-
 .../hadoop/hbase/mapreduce/TestImportTsv.java   |4 +-
 .../hadoop/hbase/mapreduce/TestSyncTable.java   |2 +-
 .../codec/prefixtree/PrefixTreeSeeker.java  |3 +-
 .../decode/PrefixTreeArrayScanner.java  |3 +-
 .../decode/PrefixTreeArraySearcher.java |7 +-
 .../codec/prefixtree/decode/PrefixTreeCell.java |4 +-
 .../prefixtree/encode/PrefixTreeEncoder.java|9 +-
 .../codec/prefixtree/row/BaseTestRowData.java   |2 +-
 .../prefixtree/row/TestPrefixTreeSearcher.java  |3 +-
 .../hbase/client/ClientSideRegionScanner.java   |3 +-
 .../hbase/client/TableSnapshotScanner.java  |3 +-
 .../hadoop/hbase/io/HalfStoreFileReader.java|   13 +-
 .../io/hfile/CompoundBloomFilterWriter.java |4 +-
 .../hadoop/hbase/io/hfile/HFileBlockIndex.java  |5 +-
 .../hbase/io/hfile/HFilePrettyPrinter.java  |3 +-
 .../hadoop/hbase/io/hfile/HFileReaderImpl.java  |   10 +-
 .../hadoop/hbase/io/hfile/HFileWriterImpl.java  |   15 +-
 .../hbase/mob/DefaultMobStoreCompactor.java |3 +-
 .../org/apache/hadoop/hbase/mob/MobUtils.java   |   11 +-
 .../hbase/protobuf/ReplicationProtbufUtil.java  |4 +-
 .../hadoop/hbase/regionserver/HRegion.java  |   44 +-
 .../hbase/regionserver/HRegionFileSystem.java   |6 +-
 .../h

[2/6] hbase git commit: HBASE-18995 Move methods that are for internal usage from CellUtil to Private util class (Ramkrishna Vasudevan)

2017-10-27 Thread stack
http://git-wip-us.apache.org/repos/asf/hbase/blob/f6c2490b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java
index a3841aa..84edf37 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityController.java
@@ -44,13 +44,13 @@ import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.hbase.AuthUtil;
 import org.apache.hadoop.hbase.Cell;
 import org.apache.hadoop.hbase.CellScanner;
-import org.apache.hadoop.hbase.CellUtil;
 import org.apache.hadoop.hbase.CoprocessorEnvironment;
 import org.apache.hadoop.hbase.DoNotRetryIOException;
 import org.apache.hadoop.hbase.HBaseInterfaceAudience;
 import org.apache.hadoop.hbase.HColumnDescriptor;
 import org.apache.hadoop.hbase.HConstants;
 import org.apache.hadoop.hbase.HTableDescriptor;
+import org.apache.hadoop.hbase.PrivateCellUtil;
 import org.apache.hadoop.hbase.MetaTableAccessor;
 import org.apache.hadoop.hbase.TableName;
 import org.apache.hadoop.hbase.Tag;
@@ -58,7 +58,6 @@ import org.apache.hadoop.hbase.TagType;
 import org.apache.hadoop.hbase.TagUtil;
 import org.apache.hadoop.hbase.client.Admin;
 import org.apache.hadoop.hbase.client.Append;
-import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
 import org.apache.hadoop.hbase.client.Delete;
 import org.apache.hadoop.hbase.client.Get;
 import org.apache.hadoop.hbase.client.Increment;
@@ -369,13 +368,13 @@ public class VisibilityController implements 
MasterCoprocessor, RegionCoprocesso
 List updatedCells = new ArrayList<>();
 for (CellScanner cellScanner = m.cellScanner(); 
cellScanner.advance();) {
   Cell cell = cellScanner.current();
-  List tags = CellUtil.getTags(cell);
+  List tags = PrivateCellUtil.getTags(cell);
   if (modifiedTagFound) {
 // Rewrite the tags by removing the modified tags.
 removeReplicationVisibilityTag(tags);
   }
   tags.addAll(visibilityTags);
-  Cell updatedCell = CellUtil.createCell(cell, tags);
+  Cell updatedCell = PrivateCellUtil.createCell(cell, tags);
   updatedCells.add(updatedCell);
 }
 m.getFamilyCellMap().clear();
@@ -428,7 +427,7 @@ public class VisibilityController implements 
MasterCoprocessor, RegionCoprocesso
 
 if (result.size() < get.getMaxVersions()) {
   // Nothing to delete
-  CellUtil.updateLatestStamp(cell, byteNow, 0);
+  PrivateCellUtil.updateLatestStamp(cell, byteNow, 0);
   return;
 }
 if (result.size() > get.getMaxVersions()) {
@@ -436,7 +435,7 @@ public class VisibilityController implements 
MasterCoprocessor, RegionCoprocesso
   + ". Results more than the max versions obtained.");
 }
 Cell getCell = result.get(get.getMaxVersions() - 1);
-CellUtil.setTimestamp(cell, getCell.getTimestamp());
+PrivateCellUtil.setTimestamp(cell, getCell.getTimestamp());
 
 // We are bypassing here because in the 
HRegion.updateDeleteLatestVersionTimeStamp we would
 // update with the current timestamp after again doing a get. As the hook 
as already determined
@@ -472,7 +471,7 @@ public class VisibilityController implements 
MasterCoprocessor, RegionCoprocesso
   // cell visiblilty tags
   // have been modified
   Tag modifiedTag = null;
-  Iterator tagsIterator = CellUtil.tagsIterator(cell);
+  Iterator tagsIterator = PrivateCellUtil.tagsIterator(cell);
   while (tagsIterator.hasNext()) {
 Tag tag = tagsIterator.next();
 if (tag.getType() == TagType.STRING_VIS_TAG_TYPE) {
@@ -484,7 +483,7 @@ public class VisibilityController implements 
MasterCoprocessor, RegionCoprocesso
   pair.setSecond(modifiedTag);
   return pair;
 }
-Iterator tagsItr = CellUtil.tagsIterator(cell);
+Iterator tagsItr = PrivateCellUtil.tagsIterator(cell);
 while (tagsItr.hasNext()) {
   if (RESERVED_VIS_TAG_TYPES.contains(tagsItr.next().getType())) {
 return pair;
@@ -514,7 +513,7 @@ public class VisibilityController implements 
MasterCoprocessor, RegionCoprocesso
 if (isSystemOrSuperUser()) {
   return true;
 }
-Iterator tagsItr = CellUtil.tagsIterator(cell);
+Iterator tagsItr = PrivateCellUtil.tagsIterator(cell);
 while (tagsItr.hasNext()) {
   if (RESERVED_VIS_TAG_TYPES.contains(tagsItr.next().getType())) {
 return false;
@@ -731,7 +730,7 @@ public class VisibilityController implements 
MasterCoprocessor, RegionCoprocesso
 
tags.addAll(this.visibilityLabel

[4/6] hbase git commit: HBASE-18995 Move methods that are for internal usage from CellUtil to Private util class (Ramkrishna Vasudevan)

2017-10-27 Thread stack
http://git-wip-us.apache.org/repos/asf/hbase/blob/f6c2490b/hbase-common/src/main/java/org/apache/hadoop/hbase/IndividualBytesFieldCell.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/IndividualBytesFieldCell.java
 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/IndividualBytesFieldCell.java
index 14e35df..a15843c 100644
--- 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/IndividualBytesFieldCell.java
+++ 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/IndividualBytesFieldCell.java
@@ -24,7 +24,6 @@ import java.io.IOException;
 import java.io.OutputStream;
 import java.nio.ByteBuffer;
 
-import org.apache.yetus.audience.InterfaceAudience;
 import org.apache.hadoop.hbase.util.ArrayUtils;
 import org.apache.hadoop.hbase.util.ByteBufferUtils;
 import org.apache.hadoop.hbase.util.Bytes;
@@ -144,7 +143,7 @@ public class IndividualBytesFieldCell implements 
ExtendedCell {
 ByteBufferUtils.putInt(out, getValueLength());
 
 // Key
-CellUtil.writeFlatKey(this, out);
+PrivateCellUtil.writeFlatKey(this, out);
 
 // Value
 out.write(getValueArray());

http://git-wip-us.apache.org/repos/asf/hbase/blob/f6c2490b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
--
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
index 6154045..42ac97d 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
@@ -1617,7 +1617,7 @@ public class KeyValue implements ExtendedCell {
  */
 @Override
 public int compare(final Cell left, final Cell right) {
-  return 
CellUtil.compareKeyIgnoresMvcc(CellComparatorImpl.META_COMPARATOR, left, right);
+  return 
PrivateCellUtil.compareKeyIgnoresMvcc(CellComparatorImpl.META_COMPARATOR, left, 
right);
 }
 
 @Override
@@ -1839,7 +1839,7 @@ public class KeyValue implements ExtendedCell {
 }
 
 public int compareOnlyKeyPortion(Cell left, Cell right) {
-  return CellUtil.compareKeyIgnoresMvcc(CellComparatorImpl.COMPARATOR, 
left, right);
+  return 
PrivateCellUtil.compareKeyIgnoresMvcc(CellComparatorImpl.COMPARATOR, left, 
right);
 }
 
 /**

http://git-wip-us.apache.org/repos/asf/hbase/blob/f6c2490b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueTestUtil.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueTestUtil.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueTestUtil.java
index c8bed3e..7467d67 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueTestUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueTestUtil.java
@@ -85,7 +85,7 @@ public class KeyValueTestUtil {
 for (Cell kv1 : kvCollection1) {
   boolean found = false;
   for (Cell kv2 : kvCollection2) {
-if (CellUtil.equalsIgnoreMvccVersion(kv1, kv2)) found = true;
+if (PrivateCellUtil.equalsIgnoreMvccVersion(kv1, kv2)) found = true;
   }
   if (!found) return false;
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/f6c2490b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java
--
diff --git 
a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java
index f1f03eb..6fd37c0 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValueUtil.java
@@ -163,7 +163,7 @@ public class KeyValueUtil {
 pos = CellUtil.copyValueTo(cell, output, pos);
 if (withTags && (cell.getTagsLength() > 0)) {
   pos = Bytes.putAsShort(output, pos, cell.getTagsLength());
-  pos = CellUtil.copyTagTo(cell, output, pos);
+  pos = PrivateCellUtil.copyTagsTo(cell, output, pos);
 }
 return pos;
   }
@@ -179,7 +179,7 @@ public class KeyValueUtil {
 int tagsLength = cell.getTagsLength();
 if (withTags && (tagsLength > 0)) {
   offset = ByteBufferUtils.putAsShort(buf, offset, tagsLength);// Tags 
length
-  offset = CellUtil.copyTagTo(cell, buf, offset);// Tags bytes
+  offset = PrivateCellUtil.copyTagsTo(cell, buf, offset);// Tags bytes
 }
 return offset;
   }

http://git-wip-us.apache.org/repos/asf/hbase/blob/f6c2490b/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java
--
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java 
b/hbase-common/src/main/java/org/apache/hadoop/hbase/TagUtil.java
index 0f010a0..a4962f4 100644
-

[1/6] hbase git commit: HBASE-18995 Move methods that are for internal usage from CellUtil to Private util class (Ramkrishna Vasudevan)

2017-10-27 Thread stack
Repository: hbase
Updated Branches:
  refs/heads/master 4dee4a854 -> f6c2490b8


http://git-wip-us.apache.org/repos/asf/hbase/blob/f6c2490b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java
--
diff --git 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java
 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java
index 6224495..e0f2986 100644
--- 
a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java
+++ 
b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestCoprocessorScanPolicy.java
@@ -262,7 +262,7 @@ public class TestCoprocessorScanPolicy {
   if (version == 0) {
 return c -> true;
   } else {
-if (row == null || !CellUtil.matchingRow(firstCell, row)) {
+if (row == null || !CellUtil.matchingRows(firstCell, row)) {
   row = CellUtil.cloneRow(firstCell);
   // reset qualifier as there is a row change
   qualifier = null;

http://git-wip-us.apache.org/repos/asf/hbase/blob/f6c2490b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftUtilities.java
--
diff --git 
a/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftUtilities.java
 
b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftUtilities.java
index 271b70c..4e15a9c 100644
--- 
a/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftUtilities.java
+++ 
b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftUtilities.java
@@ -30,6 +30,7 @@ import org.apache.commons.collections4.MapUtils;
 import org.apache.hadoop.hbase.CompareOperator;
 import org.apache.hadoop.hbase.HRegionInfo;
 import org.apache.hadoop.hbase.HRegionLocation;
+import org.apache.hadoop.hbase.PrivateCellUtil;
 import org.apache.hadoop.hbase.ServerName;
 import org.apache.yetus.audience.InterfaceAudience;
 import org.apache.hadoop.hbase.Cell;
@@ -171,7 +172,7 @@ public class ThriftUtilities {
   col.setTimestamp(kv.getTimestamp());
   col.setValue(CellUtil.cloneValue(kv));
   if (kv.getTagsLength() > 0) {
-col.setTags(CellUtil.getTagArray(kv));
+col.setTags(PrivateCellUtil.getTagsArray(kv));
   }
   columnValues.add(col);
 }



[3/6] hbase git commit: HBASE-18995 Move methods that are for internal usage from CellUtil to Private util class (Ramkrishna Vasudevan)

2017-10-27 Thread stack
http://git-wip-us.apache.org/repos/asf/hbase/blob/f6c2490b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterImpl.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterImpl.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterImpl.java
index 33cfa1d..bd98cdd 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterImpl.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterImpl.java
@@ -39,6 +39,7 @@ import org.apache.hadoop.hbase.CellComparator;
 import org.apache.hadoop.hbase.CellComparatorImpl;
 import org.apache.hadoop.hbase.CellUtil;
 import org.apache.hadoop.hbase.HConstants;
+import org.apache.hadoop.hbase.PrivateCellUtil;
 import org.apache.hadoop.hbase.KeyValueUtil;
 import org.apache.hadoop.hbase.CellComparatorImpl.MetaCellComparator;
 import org.apache.yetus.audience.InterfaceAudience;
@@ -239,7 +240,7 @@ public class HFileWriterImpl implements HFile.Writer {
   throw new IOException("Key cannot be null or empty");
 }
 if (lastCell != null) {
-  int keyComp = CellUtil.compareKeyIgnoresMvcc(comparator, lastCell, cell);
+  int keyComp = PrivateCellUtil.compareKeyIgnoresMvcc(comparator, 
lastCell, cell);
 
   if (keyComp > 0) {
 throw new IOException("Added a key not lexically larger than"
@@ -341,7 +342,7 @@ public class HFileWriterImpl implements HFile.Writer {
 int onDiskSize = blockWriter.getOnDiskSizeWithHeader();
 Cell indexEntry =
   getMidpoint(this.comparator, lastCellOfPreviousBlock, firstCellInBlock);
-
dataBlockIndexWriter.addEntry(CellUtil.getCellKeySerializedAsKeyValueKey(indexEntry),
+
dataBlockIndexWriter.addEntry(PrivateCellUtil.getCellKeySerializedAsKeyValueKey(indexEntry),
   lastDataBlockOffset, onDiskSize);
 totalUncompressedBytes += blockWriter.getUncompressedSizeWithHeader();
 if (cacheConf.shouldCacheDataOnWrite()) {
@@ -397,7 +398,7 @@ public class HFileWriterImpl implements HFile.Writer {
   }
   // If midRow is null, just return 'right'. Can't do optimization.
   if (midRow == null) return right;
-  return CellUtil.createFirstOnRow(midRow);
+  return PrivateCellUtil.createFirstOnRow(midRow);
 }
 // Rows are same. Compare on families.
 diff = comparator.compareFamilies(left, right);
@@ -419,7 +420,7 @@ public class HFileWriterImpl implements HFile.Writer {
   // If midRow is null, just return 'right'. Can't do optimization.
   if (midRow == null) return right;
   // Return new Cell where we use right row and then a mid sort family.
-  return CellUtil.createFirstOnRowFamily(right, midRow, 0, midRow.length);
+  return PrivateCellUtil.createFirstOnRowFamily(right, midRow, 0, 
midRow.length);
 }
 // Families are same. Compare on qualifiers.
 diff = comparator.compareQualifiers(left, right);
@@ -441,7 +442,7 @@ public class HFileWriterImpl implements HFile.Writer {
   // If midRow is null, just return 'right'. Can't do optimization.
   if (midRow == null) return right;
   // Return new Cell where we use right row and family and then a mid sort 
qualifier.
-  return CellUtil.createFirstOnRowCol(right, midRow, 0, midRow.length);
+  return PrivateCellUtil.createFirstOnRowCol(right, midRow, 0, 
midRow.length);
 }
 // No opportunity for optimization. Just return right key.
 return right;
@@ -738,7 +739,7 @@ public class HFileWriterImpl implements HFile.Writer {
 
 blockWriter.write(cell);
 
-totalKeyLength += CellUtil.estimatedSerializedSizeOfKey(cell);
+totalKeyLength += PrivateCellUtil.estimatedSerializedSizeOfKey(cell);
 totalValueLength += cell.getValueLength();
 
 // Are we the first key in this block?
@@ -776,7 +777,7 @@ public class HFileWriterImpl implements HFile.Writer {
 if (lastCell != null) {
   // Make a copy. The copy is stuffed into our fileinfo map. Needs a clean
   // byte buffer. Won't take a tuple.
-  byte [] lastKey = 
CellUtil.getCellKeySerializedAsKeyValueKey(this.lastCell);
+  byte [] lastKey = 
PrivateCellUtil.getCellKeySerializedAsKeyValueKey(this.lastCell);
   fileInfo.append(FileInfo.LASTKEY, lastKey, false);
 }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/f6c2490b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java
index 502a446..32552da 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mob/DefaultMobStoreCompactor.java
@@ -30,6 +30,7 @@ import

[1/2] hbase git commit: HBASE-19113 Restore dropped constants from TableInputFormatBase for compatability

2017-10-27 Thread apurtell
Repository: hbase
Updated Branches:
  refs/heads/branch-1 0697db220 -> 0d26ca7df
  refs/heads/branch-1.4 ac0383830 -> 8a5d87f41


HBASE-19113 Restore dropped constants from TableInputFormatBase for 
compatability

Signed-off-by: Sean Busbey 


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/8a5d87f4
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/8a5d87f4
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/8a5d87f4

Branch: refs/heads/branch-1.4
Commit: 8a5d87f4105e127debefc7009e0f0bba7c04c35c
Parents: ac03838
Author: Andrew Purtell 
Authored: Fri Oct 27 14:07:12 2017 -0700
Committer: Andrew Purtell 
Committed: Fri Oct 27 15:26:59 2017 -0700

--
 .../apache/hadoop/hbase/mapreduce/TableInputFormatBase.java  | 8 
 1 file changed, 8 insertions(+)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/8a5d87f4/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java
index 661b981..8d12b43 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java
@@ -128,6 +128,14 @@ extends InputFormat {
 " previous error. Please look at the previous logs lines from" +
 " the task's full log for more details.";
 
+  @Deprecated
+  /** Deprecated. No effect. */
+  public static final String INPUT_AUTOBALANCE_MAXSKEWRATIO = 
"hbase.mapreduce.input.autobalance" +
+  ".maxskewratio";
+  @Deprecated
+  /** Deprecated. No effect. */
+  public static final String TABLE_ROW_TEXTKEY = "hbase.table.row.textkey";
+
   /** Specify if we enable auto-balance to set number of mappers in M/R jobs. 
*/
   public static final String MAPREDUCE_INPUT_AUTOBALANCE = 
"hbase.mapreduce.input.autobalance";
   /** In auto-balance, we split input by ave region size, if calculated region 
size is too big, we can set it. */



[2/2] hbase git commit: HBASE-19113 Restore dropped constants from TableInputFormatBase for compatability

2017-10-27 Thread apurtell
HBASE-19113 Restore dropped constants from TableInputFormatBase for 
compatability

Signed-off-by: Sean Busbey 


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/0d26ca7d
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/0d26ca7d
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/0d26ca7d

Branch: refs/heads/branch-1
Commit: 0d26ca7dfbdf0a2b618916d17813adf6c3e273ba
Parents: 0697db2
Author: Andrew Purtell 
Authored: Fri Oct 27 14:07:12 2017 -0700
Committer: Andrew Purtell 
Committed: Fri Oct 27 15:27:32 2017 -0700

--
 .../apache/hadoop/hbase/mapreduce/TableInputFormatBase.java  | 8 
 1 file changed, 8 insertions(+)
--


http://git-wip-us.apache.org/repos/asf/hbase/blob/0d26ca7d/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java
--
diff --git 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java
 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java
index 661b981..8d12b43 100644
--- 
a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java
+++ 
b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java
@@ -128,6 +128,14 @@ extends InputFormat {
 " previous error. Please look at the previous logs lines from" +
 " the task's full log for more details.";
 
+  @Deprecated
+  /** Deprecated. No effect. */
+  public static final String INPUT_AUTOBALANCE_MAXSKEWRATIO = 
"hbase.mapreduce.input.autobalance" +
+  ".maxskewratio";
+  @Deprecated
+  /** Deprecated. No effect. */
+  public static final String TABLE_ROW_TEXTKEY = "hbase.table.row.textkey";
+
   /** Specify if we enable auto-balance to set number of mappers in M/R jobs. 
*/
   public static final String MAPREDUCE_INPUT_AUTOBALANCE = 
"hbase.mapreduce.input.autobalance";
   /** In auto-balance, we split input by ave region size, if calculated region 
size is too big, we can set it. */



[36/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/org/apache/hadoop/hbase/rest/client/RemoteHTable.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/rest/client/RemoteHTable.html 
b/apidocs/org/apache/hadoop/hbase/rest/client/RemoteHTable.html
index a56d4e7..ea9f843 100644
--- a/apidocs/org/apache/hadoop/hbase/rest/client/RemoteHTable.html
+++ b/apidocs/org/apache/hadoop/hbase/rest/client/RemoteHTable.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":42,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":42,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":42,"i51":10};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":42,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":42,"i33":10,"i34":42,"i35":10,"i36":42,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":42,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":42,"i55":10};
 var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
Methods"],8:["t4","Concrete Methods"],32:["t6","Deprecated Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -114,7 +114,7 @@ var activeTableTab = "activeTableTab";
 
 
 @InterfaceAudience.Public
-public class RemoteHTable
+public class RemoteHTable
 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
 title="class or interface in java.lang">Object
 implements Table
 HTable interface to remote tables accessed via REST 
gateway
@@ -434,64 +434,88 @@ implements 
 int
 getOperationTimeout()
-Get timeout (millisecond) of each operation for in Table 
instance.
+Deprecated. 
 
 
 
+long
+getOperationTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
 title="class or interface in 
java.util.concurrent">TimeUnit unit)
+Get timeout of each operation in Table instance.
+
+
+
 int
 getReadRpcTimeout()
-Get timeout (millisecond) of each rpc read request in this 
Table instance.
+Deprecated. 
 
 
-
+
+long
+getReadRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
 title="class or interface in 
java.util.concurrent">TimeUnit unit)
+Get timeout of each rpc read request in this Table 
instance.
+
+
+
 int
 getRpcTimeout()
 Deprecated. 
 
 
-
+
+long
+getRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
 title="class or interface in 
java.util.concurrent">TimeUnit unit)
+Get timeout of each rpc request in this Table 
instance.
+
+
+
 ResultScanner
 getScanner(byte[] family)
 Gets a scanner on the current table for the given 
family.
 
 
-
+
 ResultScanner
 getScanner(byte[] family,
   byte[] qualifier)
 Gets a scanner on the current table for the given family 
and qualifier.
 
 
-
+
 ResultScanner
 getScanner(Scan scan)
 Returns a scanner on the current table as specified by the 
Scan
  object.
 
 
-
+
 HTableDescriptor
 getTableDescriptor()
 Gets the table descriptor for 
this table.
 
 
-
+
 byte[]
 getTableName() 
 
-
+
 int
 getWriteRpcTimeout()
-Get timeout (millisecond) of each rpc write request in this 
Table instance.
+Deprecated. 
 
 
-
+
+long
+getWriteRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
 title="class or interface in 
java.util.concurrent">TimeUnit unit)
+Get timeout of each rpc write request in this Table 
instance.
+
+
+
 Result
 increment(Increment increment)
 Increments one or more columns within a single row.
 
 
-
+
 long
 incrementColumnValue(byte[] row,
 byte[] family,
@@ -500,7 +524,7 @@ implements See Table.incrementColumnValue(byte[],
 byte[], byte[], long, Durability)
 
 
-
+
 long
 incrementColumnValue(byte[] row,
 byte[] family,
@@ -510,49 +534,49 @@ implements Atomically increments a column value.
 
 
-
+
 boolean
 isAutoFlush() 
 
-
+
 void
 mutateRow(RowMutations rm)
 Performs multiple mutations atomically on a single 
row.
 
 
-
+
 void
 put(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List puts)
 Batch puts the specified data into the table.
 
 
-
+
 void
 put(Put put)
 Puts some data in the table.
 
 
-
+
 void
 setOperationTimeout(int operationTimeout)
 Set timeout (millisecond) of each operation in this Tab

[23/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/src-html/org/apache/hadoop/hbase/filter/TimestampsFilter.html
--
diff --git 
a/apidocs/src-html/org/apache/hadoop/hbase/filter/TimestampsFilter.html 
b/apidocs/src-html/org/apache/hadoop/hbase/filter/TimestampsFilter.html
index 0af43de..c6abac8 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/filter/TimestampsFilter.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/TimestampsFilter.html
@@ -116,127 +116,133 @@
 108return false;
 109  }
 110
-111  @Override
-112  public ReturnCode filterKeyValue(Cell 
v) {
-113if 
(this.timestamps.contains(v.getTimestamp())) {
-114  return ReturnCode.INCLUDE;
-115} else if (v.getTimestamp() < 
minTimeStamp) {
-116  // The remaining versions of this 
column are guaranteed
-117  // to be lesser than all of the 
other values.
-118  return ReturnCode.NEXT_COL;
-119}
-120return canHint ? 
ReturnCode.SEEK_NEXT_USING_HINT : ReturnCode.SKIP;
-121  }
-122
-123
-124  /**
-125   * Pick the next cell that the scanner 
should seek to. Since this can skip any number of cells
-126   * any of which can be a delete this 
can resurect old data.
-127   *
-128   * The method will only be used if 
canHint was set to true while creating the filter.
-129   *
-130   * @throws IOException This will never 
happen.
-131   */
-132  public Cell getNextCellHint(Cell 
currentCell) throws IOException {
-133if (!canHint) {
-134  return null;
-135}
-136
-137Long nextTimestampObject = 
timestamps.lower(currentCell.getTimestamp());
-138
-139if (nextTimestampObject == null) {
-140  // This should only happen if the 
current column's
-141  // timestamp is below the last one 
in the list.
-142  //
-143  // It should never happen as the 
filterKeyValue should return NEXT_COL
-144  // but it's always better to be 
extra safe and protect against future
-145  // behavioral changes.
-146
-147  return 
CellUtil.createLastOnRowCol(currentCell);
-148}
-149
-150// Since we know the 
nextTimestampObject isn't null here there must still be
-151// timestamps that can be included. 
Cast the Long to a long and return the
-152// a cell with the current row/cf/col 
and the next found timestamp.
-153long nextTimestamp = 
nextTimestampObject;
-154return 
CellUtil.createFirstOnRowColTS(currentCell, nextTimestamp);
-155  }
-156
-157  public static Filter 
createFilterFromArguments(ArrayList filterArguments) {
-158ArrayList timestamps = 
new ArrayList<>(filterArguments.size());
-159for (int i = 0; 
ibytes
-181   * @see #toByteArray
-182   */
-183  public static TimestampsFilter 
parseFrom(final byte[] pbBytes)
-184  throws DeserializationException {
-185FilterProtos.TimestampsFilter 
proto;
-186try {
-187  proto = 
FilterProtos.TimestampsFilter.parseFrom(pbBytes);
-188} catch 
(InvalidProtocolBufferException e) {
-189  throw new 
DeserializationException(e);
-190}
-191return new 
TimestampsFilter(proto.getTimestampsList(),
-192proto.hasCanHint() && 
proto.getCanHint());
-193  }
-194
-195  /**
-196   * @param other
-197   * @return true if and only if the 
fields of the filter that are serialized
-198   * are equal to the corresponding 
fields in other.  Used for testing.
-199   */
-200  boolean areSerializedFieldsEqual(Filter 
o) {
-201if (o == this) return true;
-202if (!(o instanceof TimestampsFilter)) 
return false;
-203
-204TimestampsFilter other = 
(TimestampsFilter)o;
-205return 
this.getTimestamps().equals(other.getTimestamps());
-206  }
-207
-208  @Override
-209  public String toString() {
-210return 
toString(MAX_LOG_TIMESTAMPS);
-211  }
-212
-213  protected String toString(int 
maxTimestamps) {
-214StringBuilder tsList = new 
StringBuilder();
-215
-216int count = 0;
-217for (Long ts : this.timestamps) {
-218  if (count >= maxTimestamps) {
-219break;
-220  }
-221  ++count;
-222  tsList.append(ts.toString());
-223  if (count < 
this.timestamps.size() && count < maxTimestamps) {
-224   

[39/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/org/apache/hadoop/hbase/filter/RandomRowFilter.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/filter/RandomRowFilter.html 
b/apidocs/org/apache/hadoop/hbase/filter/RandomRowFilter.html
index 8fed62e..bc17944 100644
--- a/apidocs/org/apache/hadoop/hbase/filter/RandomRowFilter.html
+++ b/apidocs/org/apache/hadoop/hbase/filter/RandomRowFilter.html
@@ -18,8 +18,8 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":9,"i7":10,"i8":10,"i9":10};
-var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
+var methods = 
{"i0":10,"i1":10,"i2":42,"i3":10,"i4":10,"i5":10,"i6":10,"i7":9,"i8":10,"i9":10,"i10":10};
+var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete 
Methods"],32:["t6","Deprecated Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
 var tableTab = "tableTab";
@@ -204,7 +204,7 @@ extends org.apache.hadoop.hbase.filter.FilterBase
 
 Method Summary
 
-All Methods Static Methods Instance Methods Concrete Methods 
+All Methods Static Methods Instance Methods Concrete Methods Deprecated Methods 
 
 Modifier and Type
 Method and Description
@@ -218,53 +218,59 @@ extends org.apache.hadoop.hbase.filter.FilterBase
 
 
 Filter.ReturnCode
-filterKeyValue(Cell v)
+filterCell(Cell c)
 A way to filter based on the column family, column 
qualifier and/or the column value.
 
 
 
+Filter.ReturnCode
+filterKeyValue(Cell c)
+Deprecated. 
+
+
+
 boolean
 filterRow()
 Filters that never filter by rows based on previously 
gathered state from
- Filter.filterKeyValue(Cell)
 can inherit this implementation that
+ Filter.filterCell(Cell)
 can inherit this implementation that
  never filters a row.
 
 
-
+
 boolean
 filterRowKey(Cell firstRowCell)
 Filters a row based on the row key.
 
 
-
+
 float
 getChance() 
 
-
+
 boolean
 hasFilterRow()
 Fitlers that never filter by modifying the returned List of 
Cells can
  inherit this implementation that does nothing.
 
 
-
+
 static RandomRowFilter
 parseFrom(byte[] pbBytes) 
 
-
+
 void
 reset()
 Filters that are purely stateless and do nothing in their 
reset() methods can inherit
  this null/empty implementation.
 
 
-
+
 void
 setChance(float chance)
 Set the chance that a row is included.
 
 
-
+
 byte[]
 toByteArray()
 Return length 0 byte array for Filters that don't require 
special serialization
@@ -419,7 +425,9 @@ extends org.apache.hadoop.hbase.filter.FilterBase
 
 
 filterKeyValue
-public Filter.ReturnCode filterKeyValue(Cell v)
+http://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true";
 title="class or interface in java.lang">@Deprecated
+public Filter.ReturnCode filterKeyValue(Cell c)
+Deprecated. 
 Description copied from 
class: Filter
 A way to filter based on the column family, column 
qualifier and/or the column value. Return
  code is described below. This allows filters to filter only certain number of 
columns, then
@@ -432,14 +440,48 @@ extends org.apache.hadoop.hbase.filter.FilterBase
  If your filter returns ReturnCode.NEXT_ROW, it should return
  ReturnCode.NEXT_ROW until Filter.reset()
 is called just in case the caller calls
  for the next row.
- 
+
  Concrete implementers can signal a failure condition in their code by 
throwing an
  http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException.
 
-Specified by:
+Overrides:
 filterKeyValue in
 class Filter
 Parameters:
-v - the Cell in question
+c - the Cell in question
+Returns:
+code as described below, Filter.ReturnCode.INCLUDE by default
+See Also:
+Filter.ReturnCode
+
+
+
+
+
+
+
+
+filterCell
+public Filter.ReturnCode filterCell(Cell c)
+Description copied from 
class: Filter
+A way to filter based on the column family, column 
qualifier and/or the column value. Return
+ code is described below. This allows filters to filter only certain number of 
columns, then
+ terminate without matching ever column.
+
+ If filterRowKey returns true, filterCell needs to be consistent with it.
+
+ filterCell can assume that filterRowKey has already been called for the row.
+
+ If your filter returns ReturnCode.NEXT_ROW, it should return
+ ReturnCode.NEXT_ROW until Filter.reset()
 is called just in case the caller calls
+ for the next row.
+
+ Concrete implementers can signal a failure condition in their code by 
throwing an
+ http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException.
+
+Overrides:
+filterCell in
 class Filter
+Parameters:
+c - the Cell in question
 Returns:
 code as described below
 See Also:
@@ -453,13 +495,13 @@ extends org.apache.hadoop.hbase.filter.FilterBase
 
 
 filterRow
-public bo

[06/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/devapidocs/org/apache/hadoop/hbase/filter/ColumnRangeFilter.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/filter/ColumnRangeFilter.html 
b/devapidocs/org/apache/hadoop/hbase/filter/ColumnRangeFilter.html
index fc48b1f..8a070e5 100644
--- a/devapidocs/org/apache/hadoop/hbase/filter/ColumnRangeFilter.html
+++ b/devapidocs/org/apache/hadoop/hbase/filter/ColumnRangeFilter.html
@@ -18,8 +18,8 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":9,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":9,"i12":10,"i13":10};
-var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
+var methods = 
{"i0":10,"i1":9,"i2":10,"i3":42,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":9,"i13":10,"i14":10};
+var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete 
Methods"],32:["t6","Deprecated Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
 var tableTab = "tableTab";
@@ -221,7 +221,7 @@ extends 
 Method Summary
 
-All Methods Static Methods Instance Methods Concrete Methods 
+All Methods Static Methods Instance Methods Concrete Methods Deprecated Methods 
 
 Modifier and Type
 Method and Description
@@ -238,58 +238,64 @@ extends 
 
 Filter.ReturnCode
-filterKeyValue(Cell kv)
+filterCell(Cell c)
 A way to filter based on the column family, column 
qualifier and/or the column value.
 
 
 
+Filter.ReturnCode
+filterKeyValue(Cell c)
+Deprecated. 
+
+
+
 boolean
 filterRowKey(Cell cell)
 Filters a row based on the row key.
 
 
-
+
 byte[]
 getMaxColumn() 
 
-
+
 boolean
 getMaxColumnInclusive() 
 
-
+
 byte[]
 getMinColumn() 
 
-
+
 boolean
 getMinColumnInclusive() 
 
-
+
 Cell
 getNextCellHint(Cell cell)
 Filters that are not sure which key must be next seeked to, 
can inherit
  this implementation that, by default, returns a null Cell.
 
 
-
+
 boolean
 isMaxColumnInclusive() 
 
-
+
 boolean
 isMinColumnInclusive() 
 
-
+
 static ColumnRangeFilter
 parseFrom(byte[] pbBytes) 
 
-
+
 byte[]
 toByteArray()
 Return length 0 byte array for Filters that don't require 
special serialization
 
 
-
+
 http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
 toString()
 Return filter's info for debugging and logging 
purpose.
@@ -494,7 +500,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 Description copied from 
class: Filter
 Filters a row based on the row key. If this returns true, 
the entire row will be excluded. If
- false, each KeyValue in the row will be passed to Filter.filterKeyValue(Cell)
 below.
+ false, each KeyValue in the row will be passed to Filter.filterCell(Cell)
 below.
  If Filter.filterAllRemaining()
 returns true, then Filter.filterRowKey(Cell)
 should
  also return true.
 
@@ -518,7 +524,9 @@ extends 
 
 filterKeyValue
-public Filter.ReturnCode filterKeyValue(Cell kv)
+http://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true";
 title="class or interface in java.lang">@Deprecated
+public Filter.ReturnCode filterKeyValue(Cell c)
+Deprecated. 
 Description copied from 
class: Filter
 A way to filter based on the column family, column 
qualifier and/or the column value. Return
  code is described below. This allows filters to filter only certain number of 
columns, then
@@ -531,14 +539,48 @@ extends ReturnCode.NEXT_ROW, it should return
  ReturnCode.NEXT_ROW until Filter.reset()
 is called just in case the caller calls
  for the next row.
- 
+
  Concrete implementers can signal a failure condition in their code by 
throwing an
  http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException.
 
-Specified by:
+Overrides:
 filterKeyValue in
 class Filter
 Parameters:
-kv - the Cell in question
+c - the Cell in question
+Returns:
+code as described below, Filter.ReturnCode.INCLUDE by default
+See Also:
+Filter.ReturnCode
+
+
+
+
+
+
+
+
+filterCell
+public Filter.ReturnCode filterCell(Cell c)
+Description copied from 
class: Filter
+A way to filter based on the column family, column 
qualifier and/or the column value. Return
+ code is described below. This allows filters to filter only certain number of 
columns, then
+ terminate without matching ever column.
+
+ If filterRowKey returns true, filterCell needs to be consistent with it.
+
+ filterCell can assume that filterRowKey has already been called for the row.
+
+ If your filter returns ReturnCode.NEXT_ROW, it should return
+ ReturnCode.NEXT_ROW until Filter.reset()
 is called just in case the caller calls
+ for the next row.
+
+ Concrete implementers can 

[03/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/devapidocs/org/apache/hadoop/hbase/filter/FilterWrapper.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/filter/FilterWrapper.html 
b/devapidocs/org/apache/hadoop/hbase/filter/FilterWrapper.html
index 1830d8e..02e1b24 100644
--- a/devapidocs/org/apache/hadoop/hbase/filter/FilterWrapper.html
+++ b/devapidocs/org/apache/hadoop/hbase/filter/FilterWrapper.html
@@ -18,8 +18,8 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":9,"i12":10,"i13":10,"i14":10};
-var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":42,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":9,"i13":10,"i14":10,"i15":10};
+var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete 
Methods"],32:["t6","Deprecated Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
 var tableTab = "tableTab";
@@ -203,7 +203,7 @@ extends 
 Method Summary
 
-All Methods Static Methods Instance Methods Concrete Methods 
+All Methods Static Methods Instance Methods Concrete Methods Deprecated Methods 
 
 Modifier and Type
 Method and Description
@@ -223,27 +223,33 @@ extends 
 
 Filter.ReturnCode
-filterKeyValue(Cell v)
+filterCell(Cell c)
 A way to filter based on the column family, column 
qualifier and/or the column value.
 
 
 
+Filter.ReturnCode
+filterKeyValue(Cell c)
+Deprecated. 
+
+
+
 boolean
 filterRow()
-Last chance to veto row based on previous Filter.filterKeyValue(Cell)
 calls.
+Last chance to veto row based on previous Filter.filterCell(Cell)
 calls.
 
 
-
+
 void
 filterRowCells(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List kvs)
 Chance to alter the list of Cells to be submitted.
 
 
-
+
 FilterWrapper.FilterRowRetCode
 filterRowCellsWithRet(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List kvs) 
 
-
+
 boolean
 filterRowKey(byte[] buffer,
 int offset,
@@ -251,43 +257,43 @@ extends Filters a row based on the row key.
 
 
-
+
 boolean
 filterRowKey(Cell cell)
 Filters a row based on the row key.
 
 
-
+
 Cell
 getNextCellHint(Cell currentCell)
 If the filter returns the match code SEEK_NEXT_USING_HINT, 
then it should also tell which is
  the next key it must seek to.
 
 
-
+
 boolean
 hasFilterRow()
 Primarily used to check for conflicts with scans(such as 
scans that do not read a full row at a
  time).
 
 
-
+
 boolean
 isFamilyEssential(byte[] name)
 Check that given column family is essential for filter to 
check row.
 
 
-
+
 static FilterWrapper
 parseFrom(byte[] pbBytes) 
 
-
+
 void
 reset()
 Reset the state of the filter between rows.
 
 
-
+
 byte[]
 toByteArray()
 TODO: JAVADOC
@@ -296,7 +302,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException.
 
 
-
+
 Cell
 transformCell(Cell v)
 Give the filter a chance to transform the passed 
KeyValue.
@@ -461,7 +467,7 @@ extends public boolean filterRow()
   throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 Description copied from 
class: Filter
-Last chance to veto row based on previous Filter.filterKeyValue(Cell)
 calls. The filter
+Last chance to veto row based on previous Filter.filterCell(Cell)
 calls. The filter
  needs to retain state then return a particular value for this call if they 
wish to exclude a
  row if a certain column is missing (for example).
  
@@ -515,7 +521,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 Description copied from 
class: Filter
 Filters a row based on the row key. If this returns true, 
the entire row will be excluded. If
- false, each KeyValue in the row will be passed to Filter.filterKeyValue(Cell)
 below.
+ false, each KeyValue in the row will be passed to Filter.filterCell(Cell)
 below.
  
  Concrete implementers can signal a failure condition in their code by 
throwing an
  http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException.
@@ -543,7 +549,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 Description copied from 
class: Filter
 Filters a row based on the row key. If this returns true, 
the entire row will be excluded. If
- false, each KeyValue in the row will

[47/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/org/apache/hadoop/hbase/class-use/Cell.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/class-use/Cell.html 
b/apidocs/org/apache/hadoop/hbase/class-use/Cell.html
index 8bd01ee..036d92b 100644
--- a/apidocs/org/apache/hadoop/hbase/class-use/Cell.html
+++ b/apidocs/org/apache/hadoop/hbase/class-use/Cell.html
@@ -1514,256 +1514,409 @@ Input/OutputFormats, a table indexing MapReduce job, 
and utility methods.
 
 
 Filter.ReturnCode
-MultipleColumnPrefixFilter.filterColumn(Cell cell) 
+FilterList.filterCell(Cell c) 
 
 
 Filter.ReturnCode
-ColumnPrefixFilter.filterColumn(Cell cell) 
+WhileMatchFilter.filterCell(Cell c) 
+
+
+Filter.ReturnCode
+PageFilter.filterCell(Cell ignored) 
+
+
+Filter.ReturnCode
+MultipleColumnPrefixFilter.filterCell(Cell c) 
+
+
+Filter.ReturnCode
+InclusiveStopFilter.filterCell(Cell c) 
+
+
+Filter.ReturnCode
+KeyOnlyFilter.filterCell(Cell ignored) 
 
 
 Filter.ReturnCode
-FilterList.filterKeyValue(Cell c) 
+RowFilter.filterCell(Cell v) 
 
 
 Filter.ReturnCode
-WhileMatchFilter.filterKeyValue(Cell v) 
+ColumnRangeFilter.filterCell(Cell c) 
 
 
 Filter.ReturnCode
-PageFilter.filterKeyValue(Cell ignored) 
+FamilyFilter.filterCell(Cell c) 
 
 
 Filter.ReturnCode
-MultipleColumnPrefixFilter.filterKeyValue(Cell kv) 
+RandomRowFilter.filterCell(Cell c) 
 
 
 Filter.ReturnCode
-InclusiveStopFilter.filterKeyValue(Cell v) 
+FirstKeyValueMatchingQualifiersFilter.filterCell(Cell c)
+Deprecated. 
+ 
 
 
 Filter.ReturnCode
-KeyOnlyFilter.filterKeyValue(Cell ignored) 
+SkipFilter.filterCell(Cell c) 
 
 
 Filter.ReturnCode
-RowFilter.filterKeyValue(Cell v) 
+DependentColumnFilter.filterCell(Cell c) 
 
 
 Filter.ReturnCode
-ColumnRangeFilter.filterKeyValue(Cell kv) 
+Filter.filterCell(Cell c)
+A way to filter based on the column family, column 
qualifier and/or the column value.
+
 
 
 Filter.ReturnCode
-FamilyFilter.filterKeyValue(Cell v) 
+ColumnPaginationFilter.filterCell(Cell c) 
 
 
 Filter.ReturnCode
-RandomRowFilter.filterKeyValue(Cell v) 
+ValueFilter.filterCell(Cell c) 
 
 
 Filter.ReturnCode
-FirstKeyValueMatchingQualifiersFilter.filterKeyValue(Cell v)
+ColumnCountGetFilter.filterCell(Cell c) 
+
+
+Filter.ReturnCode
+QualifierFilter.filterCell(Cell c) 
+
+
+Filter.ReturnCode
+PrefixFilter.filterCell(Cell c) 
+
+
+Filter.ReturnCode
+FuzzyRowFilter.filterCell(Cell c) 
+
+
+Filter.ReturnCode
+TimestampsFilter.filterCell(Cell c) 
+
+
+Filter.ReturnCode
+ColumnPrefixFilter.filterCell(Cell cell) 
+
+
+Filter.ReturnCode
+MultiRowRangeFilter.filterCell(Cell ignored) 
+
+
+Filter.ReturnCode
+SingleColumnValueFilter.filterCell(Cell c) 
+
+
+Filter.ReturnCode
+FirstKeyOnlyFilter.filterCell(Cell c) 
+
+
+Filter.ReturnCode
+MultipleColumnPrefixFilter.filterColumn(Cell cell) 
+
+
+Filter.ReturnCode
+ColumnPrefixFilter.filterColumn(Cell cell) 
+
+
+Filter.ReturnCode
+FilterList.filterKeyValue(Cell c)
 Deprecated. 
- 
+
+
+
+Filter.ReturnCode
+WhileMatchFilter.filterKeyValue(Cell c)
+Deprecated. 
+
 
 
 Filter.ReturnCode
-SkipFilter.filterKeyValue(Cell v) 
+PageFilter.filterKeyValue(Cell c)
+Deprecated. 
+
 
 
 Filter.ReturnCode
-DependentColumnFilter.filterKeyValue(Cell c) 
+MultipleColumnPrefixFilter.filterKeyValue(Cell c)
+Deprecated. 
+
 
 
-abstract Filter.ReturnCode
-Filter.filterKeyValue(Cell v)
-A way to filter based on the column family, column 
qualifier and/or the column value.
+Filter.ReturnCode
+InclusiveStopFilter.filterKeyValue(Cell c)
+Deprecated. 
+
+
+
+Filter.ReturnCode
+KeyOnlyFilter.filterKeyValue(Cell ignored)
+Deprecated. 
+
+
+
+Filter.ReturnCode
+RowFilter.filterKeyValue(Cell c)
+Deprecated. 
+
+
+
+Filter.ReturnCode
+ColumnRangeFilter.filterKeyValue(Cell c)
+Deprecated. 
+
+
+
+Filter.ReturnCode
+FamilyFilter.filterKeyValue(Cell c)
+Deprecated. 
+
+
+
+Filter.ReturnCode
+RandomRowFilter.filterKeyValue(Cell c)
+Deprecated. 
+
+
+
+Filter.ReturnCode
+FirstKeyValueMatchingQualifiersFilter.filterKeyValue(Cell c)
+Deprecated. 
+
+
+
+Filter.ReturnCode
+SkipFilter.filterKeyValue(Cell c)
+Deprecated. 
+
+
+
+Filter.ReturnCode
+DependentColumnFilter.filterKeyValue(Cell c)
+Deprecated. 
 
 
 
 Filter.ReturnCode
-ColumnPaginationFilter.filterKeyValue(Cell v) 
+Filter.filterKeyValue(Cell c)
+Deprecated. 
+As of release 2.0.0, this 
will be removed in HBase 3.0.0.
+ Instead use filterCell(Cell)
+
+
 
 
 Filter.ReturnCode
-ValueFilter.filterKeyValue(Cell v) 
+ColumnPaginationFilter.filterKeyValue(Cell c)
+Deprecated. 
+
 
 
 Filter.ReturnCode
-ColumnCountGetFilter.filterKeyValue(Cell v) 
+ValueFilter.filterKeyValue(Cell c)
+Deprecated. 
+
 
 
 Filter.ReturnCode
-QualifierFilter.filterKeyValue(Cell v) 
+ColumnCountGetFilter.filterKeyValue(Cell c)
+Deprecated. 
+
 
 
 Filter.ReturnCode
-PrefixFilter.filterKeyValue(Cell v) 
+QualifierFilter.filterKeyValue(Cell c)
+Deprecated. 
+
 
 
 Filter.ReturnCode
-FuzzyRowFilter.filterKeyValue(Cell c) 
+P

[21/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/book.html
--
diff --git a/book.html b/book.html
index 3c52a8e..51057fa 100644
--- a/book.html
+++ b/book.html
@@ -2555,6 +2555,21 @@ Some configurations would only appear in source code; 
the only way to identify t
 
 
 
+
+
+hbase.systemtables.compacting.memstore.type
+
+
+Description
+Determines the type of memstore to be used for system tables like META, 
namespace tables etc. By default NONE is the type and hence we use the default 
memstore for all the system tables. If we need to use compacting memstore for 
system tables then set this property to BASIC/EAGER
+
+
+Default
+NONE
+
+
+
+
 
 
 hbase.regionserver.optionalcacheflushinterval
@@ -35225,7 +35240,7 @@ The server will return cellblocks compressed using this 
same compressor as long
 
 
 Version 3.0.0-SNAPSHOT
-Last updated 2017-10-25 14:29:37 UTC
+Last updated 2017-10-27 14:29:36 UTC
 
 
 

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/bulk-loads.html
--
diff --git a/bulk-loads.html b/bulk-loads.html
index c57b7a5..f94c657 100644
--- a/bulk-loads.html
+++ b/bulk-loads.html
@@ -7,7 +7,7 @@
   
 
 
-
+
 
 Apache HBase –  
   Bulk Loads in Apache HBase (TM)
@@ -311,7 +311,7 @@ under the License. -->
 https://www.apache.org/";>The Apache Software 
Foundation.
 All rights reserved.  
 
-  Last Published: 
2017-10-25
+  Last Published: 
2017-10-27
 
 
 



[29/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html
--
diff --git 
a/apidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html 
b/apidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html
index bbcbced..b72a2e0 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html
@@ -155,489 +155,495 @@
 147return true;
 148  }
 149
-150  @Override
-151  public ReturnCode filterKeyValue(Cell 
c) {
-152final int startIndex = lastFoundIndex 
>= 0 ? lastFoundIndex : 0;
-153final int size = 
fuzzyKeysData.size();
-154for (int i = startIndex; i < size 
+ startIndex; i++) {
-155  final int index = i % size;
-156  Pair 
fuzzyData = fuzzyKeysData.get(index);
-157  // This shift is idempotent - 
always end up with 0 and -1 as mask values.
-158  for (int j = 0; j < 
fuzzyData.getSecond().length; j++) {
-159fuzzyData.getSecond()[j] 
>>= 2;
-160  }
-161  SatisfiesCode satisfiesCode =
-162  satisfies(isReversed(), 
c.getRowArray(), c.getRowOffset(), c.getRowLength(),
-163fuzzyData.getFirst(), 
fuzzyData.getSecond());
-164  if (satisfiesCode == 
SatisfiesCode.YES) {
-165lastFoundIndex = index;
-166return ReturnCode.INCLUDE;
-167  }
-168}
-169// NOT FOUND -> seek next using 
hint
-170lastFoundIndex = -1;
-171
-172return 
ReturnCode.SEEK_NEXT_USING_HINT;
-173
-174  }
-175
-176  @Override
-177  public Cell getNextCellHint(Cell 
currentCell) {
-178boolean result = 
tracker.updateTracker(currentCell);
-179if (result == false) {
-180  done = true;
-181  return null;
-182}
-183byte[] nextRowKey = 
tracker.nextRow();
-184return 
CellUtil.createFirstOnRow(nextRowKey, 0, (short) nextRowKey.length);
-185  }
-186
-187  /**
-188   * If we have multiple fuzzy keys, row 
tracker should improve overall performance. It calculates
-189   * all next rows (one per every fuzzy 
key) and put them (the fuzzy key is bundled) into a priority
-190   * queue so that the smallest row key 
always appears at queue head, which helps to decide the
-191   * "Next Cell Hint". As scanning going 
on, the number of candidate rows in the RowTracker will
-192   * remain the size of fuzzy keys until 
some of the fuzzy keys won't possibly have matches any
-193   * more.
-194   */
-195  private class RowTracker {
-196private final 
PriorityQueue>> nextRows;
-197private boolean initialized = 
false;
-198
-199RowTracker() {
-200  nextRows = new 
PriorityQueue<>(fuzzyKeysData.size(),
-201  new 
Comparator>>() {
-202@Override
-203public int 
compare(Pair> o1,
-204Pair> o2) {
-205  return isReversed()? 
Bytes.compareTo(o2.getFirst(), o1.getFirst()):
-206
Bytes.compareTo(o1.getFirst(), o2.getFirst());
-207}
-208  });
-209}
-210
-211byte[] nextRow() {
-212  if (nextRows.isEmpty()) {
-213throw new 
IllegalStateException(
-214"NextRows should not be 
empty, make sure to call nextRow() after updateTracker() return true");
-215  } else {
-216return 
nextRows.peek().getFirst();
-217  }
-218}
-219
-220boolean updateTracker(Cell 
currentCell) {
-221  if (!initialized) {
-222for (Pair 
fuzzyData : fuzzyKeysData) {
-223  updateWith(currentCell, 
fuzzyData);
-224}
-225initialized = true;
-226  } else {
-227while (!nextRows.isEmpty() 
&& !lessThan(currentCell, nextRows.peek().getFirst())) {
-228  Pair> head = nextRows.poll();
-229  Pair 
fuzzyData = head.getSecond();
-230  updateWith(currentCell, 
fuzzyData);
-231}
-232  }
-233  return !nextRows.isEmpty();
-234}
-235
-236boolean lessThan(Cell currentCell, 
byte[] nextRowKey) {
-237  int compareResult =
-238  
CellComparatorImpl.COMPARATOR.compareRows(currentCell, nextRowKey, 0, 
nextRowKey.length);
-239  return (!isReversed() && 
compareResult < 0) || (isReversed() && compareResult > 0);
+150  @Deprecated
+151  @Override
+152  public ReturnCode filterKeyValue(final 
Cell c) {
+153return filterCell(c);
+154  }
+155
+156  @Override
+157  public ReturnCode filterCell(final Cell 
c) {
+158final int startIndex = lastFoundIndex 
>= 0 ? lastFoundIndex : 0;
+159final int size = 
fuzzyKeysData.size();
+160for (int i = startIndex; i < size 
+ startIndex; i++) {
+161  final int index = i % size;
+162  Pair 
fuzz

[51/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
Published site at .


Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/00c22388
Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/00c22388
Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/00c22388

Branch: refs/heads/asf-site
Commit: 00c2238825ad0fc4ec9dbb6e6a056d4c5cf14476
Parents: 7bee80a
Author: jenkins 
Authored: Fri Oct 27 15:16:02 2017 +
Committer: jenkins 
Committed: Fri Oct 27 15:16:02 2017 +

--
 acid-semantics.html | 4 +-
 apache_hbase_reference_guide.pdf| 37649 +
 apidocs/deprecated-list.html|   374 +-
 apidocs/index-all.html  |   209 +-
 .../org/apache/hadoop/hbase/class-use/Cell.html |   283 +-
 .../hadoop/hbase/client/AsyncTableBase.html |80 +-
 apidocs/org/apache/hadoop/hbase/client/Get.html | 2 +-
 .../org/apache/hadoop/hbase/client/Query.html   | 2 +-
 .../org/apache/hadoop/hbase/client/Scan.html| 2 +-
 .../org/apache/hadoop/hbase/client/Table.html   |   294 +-
 .../hbase/filter/ColumnCountGetFilter.html  |80 +-
 .../hbase/filter/ColumnPaginationFilter.html|88 +-
 .../hadoop/hbase/filter/ColumnPrefixFilter.html |86 +-
 .../hadoop/hbase/filter/ColumnRangeFilter.html  |90 +-
 .../hadoop/hbase/filter/CompareFilter.html  | 4 +-
 .../hbase/filter/DependentColumnFilter.html |   102 +-
 .../hadoop/hbase/filter/FamilyFilter.html   |66 +-
 .../hadoop/hbase/filter/Filter.ReturnCode.html  |18 +-
 .../org/apache/hadoop/hbase/filter/Filter.html  |   114 +-
 .../apache/hadoop/hbase/filter/FilterList.html  |   125 +-
 .../hadoop/hbase/filter/FirstKeyOnlyFilter.html |80 +-
 .../FirstKeyValueMatchingQualifiersFilter.html  |58 +-
 .../hadoop/hbase/filter/FuzzyRowFilter.html |72 +-
 .../hbase/filter/InclusiveStopFilter.html   |80 +-
 .../hadoop/hbase/filter/KeyOnlyFilter.html  |75 +-
 .../hadoop/hbase/filter/LongComparator.html | 4 +-
 .../filter/MultiRowRangeFilter.RowRange.html|24 +-
 .../hbase/filter/MultiRowRangeFilter.html   |76 +-
 .../filter/MultipleColumnPrefixFilter.html  |94 +-
 .../apache/hadoop/hbase/filter/PageFilter.html  |95 +-
 .../hadoop/hbase/filter/PrefixFilter.html   |92 +-
 .../hadoop/hbase/filter/QualifierFilter.html|66 +-
 .../hadoop/hbase/filter/RandomRowFilter.html|92 +-
 .../apache/hadoop/hbase/filter/RowFilter.html   |82 +-
 .../filter/SingleColumnValueExcludeFilter.html  | 2 +-
 .../hbase/filter/SingleColumnValueFilter.html   |   116 +-
 .../apache/hadoop/hbase/filter/SkipFilter.html  |   103 +-
 .../hadoop/hbase/filter/TimestampsFilter.html   |86 +-
 .../apache/hadoop/hbase/filter/ValueFilter.html |66 +-
 .../hadoop/hbase/filter/WhileMatchFilter.html   |   109 +-
 .../filter/class-use/Filter.ReturnCode.html |   215 +-
 .../hadoop/hbase/filter/class-use/Filter.html   | 4 +-
 .../hadoop/hbase/filter/package-summary.html| 2 +-
 .../apache/hadoop/hbase/filter/package-use.html | 2 +-
 .../apache/hadoop/hbase/mapreduce/Import.html   | 8 +-
 .../hadoop/hbase/rest/client/RemoteHTable.html  |   272 +-
 .../hadoop/hbase/client/AsyncTableBase.html |   662 +-
 .../org/apache/hadoop/hbase/client/Query.html   | 2 +-
 .../org/apache/hadoop/hbase/client/Table.html   |  1462 +-
 .../hbase/filter/ColumnCountGetFilter.html  |   126 +-
 .../hbase/filter/ColumnPaginationFilter.html|   224 +-
 .../hadoop/hbase/filter/ColumnPrefixFilter.html |   198 +-
 .../hadoop/hbase/filter/ColumnRangeFilter.html  |   220 +-
 .../hbase/filter/DependentColumnFilter.html |   296 +-
 .../hadoop/hbase/filter/FamilyFilter.html   |   140 +-
 .../hadoop/hbase/filter/Filter.ReturnCode.html  |   382 +-
 .../org/apache/hadoop/hbase/filter/Filter.html  |   382 +-
 .../hbase/filter/FilterList.Operator.html   |   210 +-
 .../apache/hadoop/hbase/filter/FilterList.html  |   210 +-
 .../hadoop/hbase/filter/FirstKeyOnlyFilter.html |   136 +-
 .../FirstKeyValueMatchingQualifiersFilter.html  |   138 +-
 .../hadoop/hbase/filter/FuzzyRowFilter.html |   950 +-
 .../hbase/filter/InclusiveStopFilter.html   |   144 +-
 .../hadoop/hbase/filter/KeyOnlyFilter.html  |   654 +-
 .../hadoop/hbase/filter/LongComparator.html | 4 +-
 .../filter/MultiRowRangeFilter.RowRange.html|   760 +-
 .../hbase/filter/MultiRowRangeFilter.html   |   760 +-
 .../filter/MultipleColumnPrefixFilter.html  |   272 +-
 .../apache/hadoop/hbase/filter/PageFilter.html  |   146 +-
 .../hadoop/hbase/filter/PrefixFilter.html   |   142 +-
 .../hadoop/hbase/filter/QualifierFilter.html|   138 +-
 .../hadoop/hbase/filter/RandomRowFilter.html|   160 +-
 .../apache/hadoop/hbase/filter/RowFilter.html   |   160 +-
 .../hbase/filter

[05/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/devapidocs/org/apache/hadoop/hbase/filter/Filter.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/filter/Filter.html 
b/devapidocs/org/apache/hadoop/hbase/filter/Filter.html
index 18c153c..7a86a6c 100644
--- a/devapidocs/org/apache/hadoop/hbase/filter/Filter.html
+++ b/devapidocs/org/apache/hadoop/hbase/filter/Filter.html
@@ -18,7 +18,7 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":38,"i6":6,"i7":6,"i8":6,"i9":6,"i10":10,"i11":9,"i12":6,"i13":10,"i14":6,"i15":6};
+var methods = 
{"i0":6,"i1":6,"i2":10,"i3":42,"i4":6,"i5":6,"i6":38,"i7":6,"i8":6,"i9":6,"i10":6,"i11":10,"i12":9,"i13":6,"i14":10,"i15":6,"i16":6};
 var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],4:["t3","Abstract 
Methods"],8:["t4","Concrete Methods"],32:["t6","Deprecated Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
@@ -123,7 +123,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 reset()
 : reset the filter state before filtering a new row. 
 filterAllRemaining():
 true means row scan is over; false means keep going. 
 filterRowKey(Cell):
 true means drop this row; false means include.
-filterKeyValue(Cell):
 decides whether to include or exclude this Cell.
+filterCell(Cell):
 decides whether to include or exclude this Cell.
 See Filter.ReturnCode. 
 transformCell(Cell):
 if the Cell is included, let the filter transform the
 Cell. 
@@ -230,24 +230,33 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 
-abstract Filter.ReturnCode
-filterKeyValue(Cell v)
+Filter.ReturnCode
+filterCell(Cell c)
 A way to filter based on the column family, column 
qualifier and/or the column value.
 
 
 
+Filter.ReturnCode
+filterKeyValue(Cell c)
+Deprecated. 
+As of release 2.0.0, this 
will be removed in HBase 3.0.0.
+ Instead use filterCell(Cell)
+
+
+
+
 abstract boolean
 filterRow()
-Last chance to veto row based on previous filterKeyValue(Cell)
 calls.
+Last chance to veto row based on previous filterCell(Cell)
 calls.
 
 
-
+
 abstract void
 filterRowCells(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List kvs)
 Chance to alter the list of Cells to be submitted.
 
 
-
+
 abstract boolean
 filterRowKey(byte[] buffer,
 int offset,
@@ -258,56 +267,56 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
 
 
 
-
+
 abstract boolean
 filterRowKey(Cell firstRowCell)
 Filters a row based on the row key.
 
 
-
+
 abstract Cell
 getNextCellHint(Cell currentCell)
 If the filter returns the match code SEEK_NEXT_USING_HINT, 
then it should also tell which is
  the next key it must seek to.
 
 
-
+
 abstract boolean
 hasFilterRow()
 Primarily used to check for conflicts with scans(such as 
scans that do not read a full row at a
  time).
 
 
-
+
 abstract boolean
 isFamilyEssential(byte[] name)
 Check that given column family is essential for filter to 
check row.
 
 
-
+
 boolean
 isReversed() 
 
-
+
 static Filter
 parseFrom(byte[] pbBytes)
 Concrete implementers can signal a failure condition in 
their code by throwing an
  http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException.
 
 
-
+
 abstract void
 reset()
 Reset the state of the filter between rows.
 
 
-
+
 void
 setReversed(boolean reversed)
 alter the reversed scan flag
 
 
-
+
 abstract byte[]
 toByteArray()
 TODO: JAVADOC
@@ -316,7 +325,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?
  http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException.
 
 
-
+
 abstract Cell
 transformCell(Cell v)
 Give the filter a chance to transform the passed 
KeyValue.
@@ -410,7 +419,7 @@ public abstract boolean Deprecated. As of release 2.0.0, this will be removed in HBase 
3.0.0.
  Instead use filterRowKey(Cell)
 Filters a row based on the row key. If this returns true, 
the entire row will be excluded. If
- false, each KeyValue in the row will be passed to filterKeyValue(Cell)
 below.
+ false, each KeyValue in the row will be passed to filterCell(Cell)
 below.
  
  Concrete implementers can signal a failure condition in their code by 
throwing an
  http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException.
@@ -435,7 +444,7 @@ public abstract boolean filterRowKey(Cell firstRowCell)
   throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 Filters a row based on the row ke

[17/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html 
b/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
index ad92978..ae84eb3 100644
--- a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
+++ b/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
@@ -167,10 +167,10 @@
 
 java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true";
 title="class or interface in java.lang">Enum (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true";
 title="class or interface in java.lang">Comparable, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true";
 title="class or interface in java.io">Serializable)
 
-org.apache.hadoop.hbase.backup.BackupRestoreConstants.BackupCommand
+org.apache.hadoop.hbase.backup.BackupInfo.BackupPhase
 org.apache.hadoop.hbase.backup.BackupType
+org.apache.hadoop.hbase.backup.BackupRestoreConstants.BackupCommand
 org.apache.hadoop.hbase.backup.BackupInfo.BackupState
-org.apache.hadoop.hbase.backup.BackupInfo.BackupPhase
 
 
 



[48/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/index-all.html
--
diff --git a/apidocs/index-all.html b/apidocs/index-all.html
index 39c02e7..f39b5d1 100644
--- a/apidocs/index-all.html
+++ b/apidocs/index-all.html
@@ -5115,6 +5115,60 @@
  
 filterAllRemaining()
 - Method in class org.apache.hadoop.hbase.filter.WhileMatchFilter
  
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ColumnCountGetFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ColumnPaginationFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ColumnPrefixFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ColumnRangeFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.DependentColumnFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.FamilyFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.Filter
+
+A way to filter based on the column family, column 
qualifier and/or the column value.
+
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.FilterList
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.FirstKeyOnlyFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.FirstKeyValueMatchingQualifiersFilter
+
+Deprecated.
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.FuzzyRowFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.InclusiveStopFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.KeyOnlyFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.MultipleColumnPrefixFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.MultiRowRangeFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.PageFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.PrefixFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.QualifierFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.RandomRowFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.RowFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.SingleColumnValueFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.SkipFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.TimestampsFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ValueFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.WhileMatchFilter
+ 
 filterColumn(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ColumnPrefixFilter
  
 filterColumn(Cell)
 - Method in class org.apache.hadoop.hbase.filter.MultipleColumnPrefixFilter
@@ -5122,59 +5176,108 @@
 filterIfMissing
 - Variable in class org.apache.hadoop.hbase.filter.SingleColumnValueFilter
  
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ColumnCountGetFilter
- 
+
+Deprecated.
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ColumnPaginationFilter
- 
+
+Deprecated.
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ColumnPrefixFilter
- 
+
+Deprecated.
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ColumnRangeFilter
- 
+
+Deprecated.
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.DependentColumnFilter
- 
+
+Deprecated.
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.FamilyFilter
- 
+
+Deprecated.
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.Filter
 
-A way to filter based on the column family, column 
qualifier and/or the column value.
+Deprecated.
+As of release 2.0.0, this 
will be removed in HBase 3.0.0.
+ Instead use filterCell(Cell)
+
 
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.FilterList
- 
+
+Deprecated.
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.FirstKeyOnlyFilter
- 
+
+Deprecated.
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.FirstKeyValueMatchingQualifiersFilter
 
 Deprecated.
- 
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.FuzzyRowFilter
- 
+
+Deprecated.
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.InclusiveStopFilter
- 
+
+Deprecated.
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.KeyOnlyFilter
- 
+
+Deprecated.
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.MultipleColumnPrefixFilter
- 
+
+Deprecated.
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.MultiRowRangeFilter
- 
+
+Deprecated.
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.PageFilter
- 
+
+Depre

[30/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html
--
diff --git 
a/apidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html 
b/apidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html
index f6b3fec..b2d188d 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.html
@@ -59,73 +59,79 @@
 051return false;
 052  }
 053
-054  @Override
-055  public ReturnCode filterKeyValue(Cell 
v) {
-056if(foundKV) return 
ReturnCode.NEXT_ROW;
-057foundKV = true;
-058return ReturnCode.INCLUDE;
-059  }
-060
-061  public static Filter 
createFilterFromArguments(ArrayList filterArguments) {
-062
Preconditions.checkArgument(filterArguments.isEmpty(),
-063"Expected 
0 but got: %s", filterArguments.size());
-064return new FirstKeyOnlyFilter();
+054  @Deprecated
+055  @Override
+056  public ReturnCode filterKeyValue(final 
Cell c) {
+057return filterCell(c);
+058  }
+059
+060  @Override
+061  public ReturnCode filterCell(final Cell 
c) {
+062if(foundKV) return 
ReturnCode.NEXT_ROW;
+063foundKV = true;
+064return ReturnCode.INCLUDE;
 065  }
 066
-067  /**
-068   * @return true if first KV has been 
found.
-069   */
-070  protected boolean hasFoundKV() {
-071return this.foundKV;
-072  }
-073
-074  /**
-075   *
-076   * @param value update {@link #foundKV} 
flag with value.
-077   */
-078  protected void setFoundKV(boolean 
value) {
-079this.foundKV = value;
-080  }
-081
-082  /**
-083   * @return The filter serialized using 
pb
-084   */
-085  public byte [] toByteArray() {
-086
FilterProtos.FirstKeyOnlyFilter.Builder builder =
-087  
FilterProtos.FirstKeyOnlyFilter.newBuilder();
-088return 
builder.build().toByteArray();
-089  }
-090
-091  /**
-092   * @param pbBytes A pb serialized 
{@link FirstKeyOnlyFilter} instance
-093   * @return An instance of {@link 
FirstKeyOnlyFilter} made from bytes
-094   * @throws 
org.apache.hadoop.hbase.exceptions.DeserializationException
-095   * @see #toByteArray
-096   */
-097  public static FirstKeyOnlyFilter 
parseFrom(final byte [] pbBytes)
-098  throws DeserializationException {
-099// There is nothing to deserialize.  
Why do this at all?
-100try {
-101  
FilterProtos.FirstKeyOnlyFilter.parseFrom(pbBytes);
-102} catch 
(InvalidProtocolBufferException e) {
-103  throw new 
DeserializationException(e);
-104}
-105// Just return a new instance.
-106return new FirstKeyOnlyFilter();
-107  }
-108
-109  /**
-110   * @param other
-111   * @return true if and only if the 
fields of the filter that are serialized
-112   * are equal to the corresponding 
fields in other.  Used for testing.
-113   */
-114  boolean areSerializedFieldsEqual(Filter 
o) {
-115if (o == this) return true;
-116if (!(o instanceof 
FirstKeyOnlyFilter)) return false;
-117
-118return true;
-119  }
-120}
+067  public static Filter 
createFilterFromArguments(ArrayList filterArguments) {
+068
Preconditions.checkArgument(filterArguments.isEmpty(),
+069"Expected 
0 but got: %s", filterArguments.size());
+070return new FirstKeyOnlyFilter();
+071  }
+072
+073  /**
+074   * @return true if first KV has been 
found.
+075   */
+076  protected boolean hasFoundKV() {
+077return this.foundKV;
+078  }
+079
+080  /**
+081   *
+082   * @param value update {@link #foundKV} 
flag with value.
+083   */
+084  protected void setFoundKV(boolean 
value) {
+085this.foundKV = value;
+086  }
+087
+088  /**
+089   * @return The filter serialized using 
pb
+090   */
+091  public byte [] toByteArray() {
+092
FilterProtos.FirstKeyOnlyFilter.Builder builder =
+093  
FilterProtos.FirstKeyOnlyFilter.newBuilder();
+094return 
builder.build().toByteArray();
+095  }
+096
+097  /**
+098   * @param pbBytes A pb serialized 
{@link FirstKeyOnlyFilter} instance
+099   * @return An instance of {@link 
FirstKeyOnlyFilter} made from bytes
+100   * @throws 
org.apache.hadoop.hbase.exceptions.DeserializationException
+101   * @see #toByteArray
+102   */
+103  public static FirstKeyOnlyFilter 
parseFrom(final byte [] pbBytes)
+104  throws DeserializationException {
+105// There is nothing to deserialize.  
Why do this at all?
+106try {
+107  
FilterProtos.FirstKeyOnlyFilter.parseFrom(pbBytes);
+108} catch 
(InvalidProtocolBufferException e) {
+109  throw new 
DeserializationException(e);
+110}
+111// Just return a new instance.
+112return new FirstKeyOnlyFilter();
+113  }
+114
+115  /**
+116   * @param o the other filter to compare 
with
+117   * @return true if and only if the 
fields of the filter that are serialized
+118   * are

[02/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/devapidocs/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html 
b/devapidocs/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html
index 65ddd82..9d6aea3 100644
--- a/devapidocs/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html
+++ b/devapidocs/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html
@@ -18,8 +18,8 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":9,"i5":9,"i6":9,"i7":10,"i8":9,"i9":10,"i10":10,"i11":9,"i12":9,"i13":9,"i14":9,"i15":10,"i16":10,"i17":9};
-var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":42,"i4":10,"i5":9,"i6":9,"i7":9,"i8":10,"i9":9,"i10":10,"i11":10,"i12":9,"i13":9,"i14":9,"i15":9,"i16":10,"i17":10,"i18":9};
+var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete 
Methods"],32:["t6","Deprecated Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
 var tableTab = "tableTab";
@@ -252,7 +252,7 @@ extends 
 Method Summary
 
-All Methods Static Methods Instance Methods Concrete Methods 
+All Methods Static Methods Instance Methods Concrete Methods Deprecated Methods 
 
 Modifier and Type
 Method and Description
@@ -272,25 +272,31 @@ extends 
 
 Filter.ReturnCode
-filterKeyValue(Cell c)
+filterCell(Cell c)
 A way to filter based on the column family, column 
qualifier and/or the column value.
 
 
 
+Filter.ReturnCode
+filterKeyValue(Cell c)
+Deprecated. 
+
+
+
 Cell
 getNextCellHint(Cell currentCell)
 Filters that are not sure which key must be next seeked to, 
can inherit
  this implementation that, by default, returns a null Cell.
 
 
-
+
 (package private) static byte[]
 getNextForFuzzyRule(boolean reverse,
byte[] row,
byte[] fuzzyKeyBytes,
byte[] fuzzyKeyMeta) 
 
-
+
 (package private) static byte[]
 getNextForFuzzyRule(boolean reverse,
byte[] row,
@@ -299,39 +305,39 @@ extends  
 
-
+
 (package private) static byte[]
 getNextForFuzzyRule(byte[] row,
byte[] fuzzyKeyBytes,
byte[] fuzzyKeyMeta) 
 
-
+
 private boolean
 isPreprocessedMask(byte[] mask) 
 
-
+
 static FuzzyRowFilter
 parseFrom(byte[] pbBytes) 
 
-
+
 private byte[]
 preprocessMask(byte[] mask)
 We need to preprocess mask array, as since we treat 2's as 
unfixed positions and -1 (0xff) as
  fixed positions
 
 
-
+
 private void
 preprocessSearchKey(Pair p) 
 
-
+
 (package private) static FuzzyRowFilter.SatisfiesCode
 satisfies(boolean reverse,
  byte[] row,
  byte[] fuzzyKeyBytes,
  byte[] fuzzyKeyMeta) 
 
-
+
 (package private) static FuzzyRowFilter.SatisfiesCode
 satisfies(boolean reverse,
  byte[] row,
@@ -340,13 +346,13 @@ extends  
 
-
+
 (package private) static FuzzyRowFilter.SatisfiesCode
 satisfies(byte[] row,
  byte[] fuzzyKeyBytes,
  byte[] fuzzyKeyMeta) 
 
-
+
 (package private) static FuzzyRowFilter.SatisfiesCode
 satisfiesNoUnsafe(boolean reverse,
  byte[] row,
@@ -355,19 +361,19 @@ extends  
 
-
+
 byte[]
 toByteArray()
 Return length 0 byte array for Filters that don't require 
special serialization
 
 
-
+
 http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
 toString()
 Return filter's info for debugging and logging 
purpose.
 
 
-
+
 private static byte[]
 trimTrailingZeroes(byte[] result,
   byte[] fuzzyKeyMeta,
@@ -529,7 +535,9 @@ extends 
 
 filterKeyValue
-public Filter.ReturnCode filterKeyValue(Cell c)
+http://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true";
 title="class or interface in java.lang">@Deprecated
+public Filter.ReturnCode filterKeyValue(Cell c)
+Deprecated. 
 Description copied from 
class: Filter
 A way to filter based on the column family, column 
qualifier and/or the column value. Return
  code is described below. This allows filters to filter only certain number of 
columns, then
@@ -542,15 +550,49 @@ extends ReturnCode.NEXT_ROW, it should return
  ReturnCode.NEXT_ROW until Filter.reset()
 is called just in case the caller calls
  for the next row.
- 
+
  Concrete implementers can signal a failure condition in their code by 
throwing an
  http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException.
 
-Specified by:
+Overrides:
 filterKeyValue in
 class Filter
 Parameters:
 c - the Cell in question
 Returns:
+code as described below, Filter.ReturnCode.INCLUDE by default
+See Also:
+Filter.ReturnCode
+
+
+
+
+
+
+
+
+filterCell
+public Filter.ReturnCode filt

[33/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/src-html/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.html
--
diff --git 
a/apidocs/src-html/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.html 
b/apidocs/src-html/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.html
index 52745c5..e098e72 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.html
@@ -70,69 +70,75 @@
 062return this.count > this.limit;
 063  }
 064
-065  @Override
-066  public ReturnCode filterKeyValue(Cell 
v) {
-067this.count++;
-068return filterAllRemaining() ? 
ReturnCode.NEXT_COL : ReturnCode.INCLUDE_AND_NEXT_COL;
+065  @Deprecated
+066  @Override
+067  public ReturnCode filterKeyValue(final 
Cell c) {
+068return filterCell(c);
 069  }
 070
 071  @Override
-072  public void reset() {
-073this.count = 0;
-074  }
-075
-076  public static Filter 
createFilterFromArguments(ArrayList filterArguments) {
-077
Preconditions.checkArgument(filterArguments.size() == 1,
-078"Expected 
1 but got: %s", filterArguments.size());
-079int limit = 
ParseFilter.convertByteArrayToInt(filterArguments.get(0));
-080return new 
ColumnCountGetFilter(limit);
-081  }
-082
-083  /**
-084   * @return The filter serialized using 
pb
-085   */
-086  public byte [] toByteArray() {
-087
FilterProtos.ColumnCountGetFilter.Builder builder =
-088  
FilterProtos.ColumnCountGetFilter.newBuilder();
-089builder.setLimit(this.limit);
-090return 
builder.build().toByteArray();
-091  }
-092
-093  /**
-094   * @param pbBytes A pb serialized 
{@link ColumnCountGetFilter} instance
-095   * @return An instance of {@link 
ColumnCountGetFilter} made from bytes
-096   * @throws 
org.apache.hadoop.hbase.exceptions.DeserializationException
-097   * @see #toByteArray
-098   */
-099  public static ColumnCountGetFilter 
parseFrom(final byte [] pbBytes)
-100  throws DeserializationException {
-101FilterProtos.ColumnCountGetFilter 
proto;
-102try {
-103  proto = 
FilterProtos.ColumnCountGetFilter.parseFrom(pbBytes);
-104} catch 
(InvalidProtocolBufferException e) {
-105  throw new 
DeserializationException(e);
-106}
-107return new 
ColumnCountGetFilter(proto.getLimit());
-108  }
-109
-110  /**
-111   * @param other
-112   * @return true if and only if the 
fields of the filter that are serialized
-113   * are equal to the corresponding 
fields in other.  Used for testing.
-114   */
-115  boolean areSerializedFieldsEqual(Filter 
o) {
-116if (o == this) return true;
-117if (!(o instanceof 
ColumnCountGetFilter)) return false;
-118
-119ColumnCountGetFilter other = 
(ColumnCountGetFilter)o;
-120return this.getLimit() == 
other.getLimit();
-121  }
-122
-123  @Override
-124  public String toString() {
-125return 
this.getClass().getSimpleName() + " " + this.limit;
-126  }
-127}
+072  public ReturnCode filterCell(final Cell 
c) {
+073this.count++;
+074return filterAllRemaining() ? 
ReturnCode.NEXT_COL : ReturnCode.INCLUDE_AND_NEXT_COL;
+075  }
+076
+077  @Override
+078  public void reset() {
+079this.count = 0;
+080  }
+081
+082  public static Filter 
createFilterFromArguments(ArrayList filterArguments) {
+083
Preconditions.checkArgument(filterArguments.size() == 1,
+084"Expected 
1 but got: %s", filterArguments.size());
+085int limit = 
ParseFilter.convertByteArrayToInt(filterArguments.get(0));
+086return new 
ColumnCountGetFilter(limit);
+087  }
+088
+089  /**
+090   * @return The filter serialized using 
pb
+091   */
+092  public byte [] toByteArray() {
+093
FilterProtos.ColumnCountGetFilter.Builder builder =
+094  
FilterProtos.ColumnCountGetFilter.newBuilder();
+095builder.setLimit(this.limit);
+096return 
builder.build().toByteArray();
+097  }
+098
+099  /**
+100   * @param pbBytes A pb serialized 
{@link ColumnCountGetFilter} instance
+101   * @return An instance of {@link 
ColumnCountGetFilter} made from bytes
+102   * @throws 
org.apache.hadoop.hbase.exceptions.DeserializationException
+103   * @see #toByteArray
+104   */
+105  public static ColumnCountGetFilter 
parseFrom(final byte [] pbBytes)
+106  throws DeserializationException {
+107FilterProtos.ColumnCountGetFilter 
proto;
+108try {
+109  proto = 
FilterProtos.ColumnCountGetFilter.parseFrom(pbBytes);
+110} catch 
(InvalidProtocolBufferException e) {
+111  throw new 
DeserializationException(e);
+112}
+113return new 
ColumnCountGetFilter(proto.getLimit());
+114  }
+115
+116  /**
+117   * @param o the other filter to compare 
with
+118   * @return true if and only if the 
fields of the filter that are serialized
+119   * are equal to the corresponding 
fields

[27/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/src-html/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.RowRange.html
--
diff --git 
a/apidocs/src-html/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.RowRange.html
 
b/apidocs/src-html/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.RowRange.html
index d780353..1554d23 100644
--- 
a/apidocs/src-html/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.RowRange.html
+++ 
b/apidocs/src-html/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.RowRange.html
@@ -132,387 +132,393 @@
 124return false;
 125  }
 126
-127  @Override
-128  public ReturnCode filterKeyValue(Cell 
ignored) {
-129return currentReturnCode;
-130  }
-131
-132  @Override
-133  public Cell getNextCellHint(Cell 
currentKV) {
-134// skip to the next range's start 
row
-135return 
CellUtil.createFirstOnRow(range.startRow, 0,
-136(short) range.startRow.length);
-137  }
-138
-139  /**
-140   * @return The filter serialized using 
pb
-141   */
-142  public byte[] toByteArray() {
-143
FilterProtos.MultiRowRangeFilter.Builder builder = 
FilterProtos.MultiRowRangeFilter
-144.newBuilder();
-145for (RowRange range : rangeList) {
-146  if (range != null) {
-147FilterProtos.RowRange.Builder 
rangebuilder = FilterProtos.RowRange.newBuilder();
-148if (range.startRow != null)
-149  
rangebuilder.setStartRow(UnsafeByteOperations.unsafeWrap(range.startRow));
-150
rangebuilder.setStartRowInclusive(range.startRowInclusive);
-151if (range.stopRow != null)
-152  
rangebuilder.setStopRow(UnsafeByteOperations.unsafeWrap(range.stopRow));
-153
rangebuilder.setStopRowInclusive(range.stopRowInclusive);
-154
builder.addRowRangeList(rangebuilder.build());
-155  }
-156}
-157return 
builder.build().toByteArray();
-158  }
-159
-160  /**
-161   * @param pbBytes A pb serialized 
instance
-162   * @return An instance of 
MultiRowRangeFilter
-163   * @throws 
org.apache.hadoop.hbase.exceptions.DeserializationException
-164   */
-165  public static MultiRowRangeFilter 
parseFrom(final byte[] pbBytes)
-166  throws DeserializationException {
-167FilterProtos.MultiRowRangeFilter 
proto;
-168try {
-169  proto = 
FilterProtos.MultiRowRangeFilter.parseFrom(pbBytes);
-170} catch 
(InvalidProtocolBufferException e) {
-171  throw new 
DeserializationException(e);
-172}
-173int length = 
proto.getRowRangeListCount();
-174List 
rangeProtos = proto.getRowRangeListList();
-175List rangeList = new 
ArrayList<>(length);
-176for (FilterProtos.RowRange rangeProto 
: rangeProtos) {
-177  RowRange range = new 
RowRange(rangeProto.hasStartRow() ? rangeProto.getStartRow()
-178  .toByteArray() : null, 
rangeProto.getStartRowInclusive(), rangeProto.hasStopRow() ?
-179  
rangeProto.getStopRow().toByteArray() : null, 
rangeProto.getStopRowInclusive());
-180  rangeList.add(range);
-181}
-182return new 
MultiRowRangeFilter(rangeList);
-183  }
-184
-185  /**
-186   * @param o the filter to compare
-187   * @return true if and only if the 
fields of the filter that are serialized are equal to the
-188   * corresponding fields in 
other. Used for testing.
-189   */
-190  boolean areSerializedFieldsEqual(Filter 
o) {
-191if (o == this)
-192  return true;
-193if (!(o instanceof 
MultiRowRangeFilter))
-194  return false;
-195
-196MultiRowRangeFilter other = 
(MultiRowRangeFilter) o;
-197if (this.rangeList.size() != 
other.rangeList.size())
-198  return false;
-199for (int i = 0; i < 
rangeList.size(); ++i) {
-200  RowRange thisRange = 
this.rangeList.get(i);
-201  RowRange otherRange = 
other.rangeList.get(i);
-202  if 
(!(Bytes.equals(thisRange.startRow, otherRange.startRow) && 
Bytes.equals(
-203  thisRange.stopRow, 
otherRange.stopRow) && (thisRange.startRowInclusive ==
-204  otherRange.startRowInclusive) 
&& (thisRange.stopRowInclusive ==
-205  otherRange.stopRowInclusive))) 
{
-206return false;
-207  }
-208}
-209return true;
-210  }
-211
-212  /**
-213   * calculate the position where the row 
key in the ranges list.
-214   *
-215   * @param rowKey the row key to 
calculate
-216   * @return index the position of the 
row key
-217   */
-218  private int getNextRangeIndex(byte[] 
rowKey) {
-219RowRange temp = new RowRange(rowKey, 
true, null, true);
-220int index = 
Collections.binarySearch(rangeList, temp);
-221if (index < 0) {
-222  int insertionPosition = -index - 
1;
-223  // check if the row key in the 
range before the insertion position
-224  if (insertionPosition != 0 
&& rangeList.get(insertionPosition - 1).contains(rowKey)) {
-225return insertionPosition - 1;
-226  }
-227  // check if the ro

[42/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/org/apache/hadoop/hbase/filter/FilterList.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/filter/FilterList.html 
b/apidocs/org/apache/hadoop/hbase/filter/FilterList.html
index face26a..a1b8f66 100644
--- a/apidocs/org/apache/hadoop/hbase/filter/FilterList.html
+++ b/apidocs/org/apache/hadoop/hbase/filter/FilterList.html
@@ -18,8 +18,8 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":9,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10};
-var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
+var methods = 
{"i0":10,"i1":10,"i2":10,"i3":10,"i4":42,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":9,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10};
+var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete 
Methods"],32:["t6","Deprecated Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
 var tableTab = "tableTab";
@@ -127,7 +127,7 @@ extends org.apache.hadoop.hbase.filter.FilterBase
  FilterList.Operator.MUST_PASS_ONE
 (OR). Since you can use Filter Lists as children of
  Filter Lists, you can create a hierarchy of filters to be evaluated. 
  FilterList.Operator.MUST_PASS_ALL
 evaluates lazily: evaluation stops as soon as one filter does not
- include the KeyValue. 
+ include the Cell. 
  FilterList.Operator.MUST_PASS_ONE
 evaluates non-lazily: all filters are always evaluated. 
  Defaults to FilterList.Operator.MUST_PASS_ALL.
 
@@ -227,7 +227,7 @@ extends org.apache.hadoop.hbase.filter.FilterBase
 
 Method Summary
 
-All Methods Static Methods Instance Methods Concrete Methods 
+All Methods Static Methods Instance Methods Concrete Methods Deprecated Methods 
 
 Modifier and Type
 Method and Description
@@ -251,26 +251,32 @@ extends org.apache.hadoop.hbase.filter.FilterBase
 
 
 Filter.ReturnCode
-filterKeyValue(Cell c)
+filterCell(Cell c)
 A way to filter based on the column family, column 
qualifier and/or the column value.
 
 
 
+Filter.ReturnCode
+filterKeyValue(Cell c)
+Deprecated. 
+
+
+
 boolean
 filterRow()
 Filters that never filter by rows based on previously 
gathered state from
- Filter.filterKeyValue(Cell)
 can inherit this implementation that
+ Filter.filterCell(Cell)
 can inherit this implementation that
  never filters a row.
 
 
-
+
 void
 filterRowCells(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List cells)
 Filters that never filter by modifying the returned List of 
Cells can inherit this
  implementation that does nothing.
 
 
-
+
 boolean
 filterRowKey(byte[] rowKey,
 int offset,
@@ -279,82 +285,82 @@ extends org.apache.hadoop.hbase.filter.FilterBase
  never filters anything.
 
 
-
+
 boolean
 filterRowKey(Cell firstRowCell)
 Filters a row based on the row key.
 
 
-
+
 http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List
 getFilters()
 Get the filters.
 
 
-
+
 Cell
 getNextCellHint(Cell currentCell)
 Filters that are not sure which key must be next seeked to, 
can inherit
  this implementation that, by default, returns a null Cell.
 
 
-
+
 FilterList.Operator
 getOperator()
 Get the operator.
 
 
-
+
 boolean
 hasFilterRow()
 Fitlers that never filter by modifying the returned List of 
Cells can
  inherit this implementation that does nothing.
 
 
-
+
 boolean
 isFamilyEssential(byte[] name)
 By default, we require all scan's column families to be 
present.
 
 
-
+
 boolean
 isReversed() 
 
-
+
 static FilterList
 parseFrom(byte[] pbBytes) 
 
-
+
 void
 reset()
 Filters that are purely stateless and do nothing in their 
reset() methods can inherit
  this null/empty implementation.
 
 
-
+
 void
 setReversed(boolean reversed)
 alter the reversed scan flag
 
 
-
+
 int
 size() 
 
-
+
 byte[]
 toByteArray()
 Return length 0 byte array for Filters that don't require 
special serialization
 
 
-
+
 http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
 toString()
 Return filter's info for debugging and logging 
purpose.
 
 
-
+
 Cell
 transformCell(Cell c)
 By default no transformation takes place
@@ -574,7 +580,7 @@ extends org.apache.hadoop.hbase.filter.FilterBase
  never filters anything. (ie: returns false).
 
  Filters a row based on the row key. If this returns true, the entire row will 
be excluded. If
- false, each KeyValue in the row will be passed to Filter.filterKeyValue(Cell)
 below.
+ false, each KeyValue in the row will be passed to Filter.filterCell(Cell)
 below.
  
  Con

[07/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/devapidocs/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.html 
b/devapidocs/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.html
index 7a20bbf..bd43512 100644
--- a/devapidocs/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.html
+++ b/devapidocs/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.html
@@ -18,8 +18,8 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":10,"i1":9,"i2":10,"i3":10,"i4":10,"i5":10,"i6":9,"i7":10,"i8":10,"i9":10};
-var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
+var methods = 
{"i0":10,"i1":9,"i2":10,"i3":10,"i4":42,"i5":10,"i6":10,"i7":9,"i8":10,"i9":10,"i10":10};
+var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete 
Methods"],32:["t6","Deprecated Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
 var tableTab = "tableTab";
@@ -201,7 +201,7 @@ extends 
 Method Summary
 
-All Methods Static Methods Instance Methods Concrete Methods 
+All Methods Static Methods Instance Methods Concrete Methods Deprecated Methods 
 
 Modifier and Type
 Method and Description
@@ -225,38 +225,44 @@ extends 
 
 Filter.ReturnCode
-filterKeyValue(Cell v)
+filterCell(Cell c)
 A way to filter based on the column family, column 
qualifier and/or the column value.
 
 
 
+Filter.ReturnCode
+filterKeyValue(Cell c)
+Deprecated. 
+
+
+
 boolean
 filterRowKey(Cell cell)
 Filters a row based on the row key.
 
 
-
+
 int
 getLimit() 
 
-
+
 static ColumnCountGetFilter
 parseFrom(byte[] pbBytes) 
 
-
+
 void
 reset()
 Filters that are purely stateless and do nothing in their 
reset() methods can inherit
  this null/empty implementation.
 
 
-
+
 byte[]
 toByteArray()
 Return length 0 byte array for Filters that don't require 
special serialization
 
 
-
+
 http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
 toString()
 Return filter's info for debugging and logging 
purpose.
@@ -360,7 +366,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 Description copied from 
class: Filter
 Filters a row based on the row key. If this returns true, 
the entire row will be excluded. If
- false, each KeyValue in the row will be passed to Filter.filterKeyValue(Cell)
 below.
+ false, each KeyValue in the row will be passed to Filter.filterCell(Cell)
 below.
  If Filter.filterAllRemaining()
 returns true, then Filter.filterRowKey(Cell)
 should
  also return true.
 
@@ -407,7 +413,9 @@ extends 
 
 filterKeyValue
-public Filter.ReturnCode filterKeyValue(Cell v)
+http://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true";
 title="class or interface in java.lang">@Deprecated
+public Filter.ReturnCode filterKeyValue(Cell c)
+Deprecated. 
 Description copied from 
class: Filter
 A way to filter based on the column family, column 
qualifier and/or the column value. Return
  code is described below. This allows filters to filter only certain number of 
columns, then
@@ -420,14 +428,48 @@ extends ReturnCode.NEXT_ROW, it should return
  ReturnCode.NEXT_ROW until Filter.reset()
 is called just in case the caller calls
  for the next row.
- 
+
  Concrete implementers can signal a failure condition in their code by 
throwing an
  http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException.
 
-Specified by:
+Overrides:
 filterKeyValue in
 class Filter
 Parameters:
-v - the Cell in question
+c - the Cell in question
+Returns:
+code as described below, Filter.ReturnCode.INCLUDE by default
+See Also:
+Filter.ReturnCode
+
+
+
+
+
+
+
+
+filterCell
+public Filter.ReturnCode filterCell(Cell c)
+Description copied from 
class: Filter
+A way to filter based on the column family, column 
qualifier and/or the column value. Return
+ code is described below. This allows filters to filter only certain number of 
columns, then
+ terminate without matching ever column.
+
+ If filterRowKey returns true, filterCell needs to be consistent with it.
+
+ filterCell can assume that filterRowKey has already been called for the row.
+
+ If your filter returns ReturnCode.NEXT_ROW, it should return
+ ReturnCode.NEXT_ROW until Filter.reset()
 is called just in case the caller calls
+ for the next row.
+
+ Concrete implementers can signal a failure condition in their code by 
throwing an
+ http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException.
+
+Overrides:
+filterCell in
 class Filter
+Parameters:
+c - the Cell in question
 Returns:

[34/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html
--
diff --git a/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html 
b/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html
index 837b9f9..c1253d9 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/client/Table.html
@@ -30,720 +30,762 @@
 022import java.io.IOException;
 023import java.util.List;
 024import java.util.Map;
-025
-026import 
org.apache.hadoop.conf.Configuration;
-027import 
org.apache.hadoop.hbase.CompareOperator;
-028import 
org.apache.hadoop.hbase.HTableDescriptor;
-029import 
org.apache.hadoop.hbase.TableName;
-030import 
org.apache.yetus.audience.InterfaceAudience;
-031import 
org.apache.hadoop.hbase.client.coprocessor.Batch;
-032import 
org.apache.hadoop.hbase.filter.CompareFilter;
-033import 
org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
-034
-035import com.google.protobuf.Descriptors;
-036import com.google.protobuf.Message;
-037import com.google.protobuf.Service;
-038import 
com.google.protobuf.ServiceException;
-039
-040/**
-041 * Used to communicate with a single 
HBase table.
-042 * Obtain an instance from a {@link 
Connection} and call {@link #close()} afterwards.
-043 *
-044 * 

Table can be used to get, put, delete or scan data from a table. -045 * @see ConnectionFactory -046 * @see Connection -047 * @see Admin -048 * @see RegionLocator -049 * @since 0.99.0 -050 */ -051@InterfaceAudience.Public -052public interface Table extends Closeable { -053 /** -054 * Gets the fully qualified table name instance of this table. -055 */ -056 TableName getName(); -057 -058 /** -059 * Returns the {@link org.apache.hadoop.conf.Configuration} object used by this instance. -060 *

-061 * The reference returned is not a copy, so any change made to it will -062 * affect this instance. -063 */ -064 Configuration getConfiguration(); -065 -066 /** -067 * Gets the {@link org.apache.hadoop.hbase.HTableDescriptor table descriptor} for this table. -068 * @throws java.io.IOException if a remote or network exception occurs. -069 * @deprecated since 2.0 version and will be removed in 3.0 version. -070 * use {@link #getDescriptor()} -071 */ -072 @Deprecated -073 HTableDescriptor getTableDescriptor() throws IOException; -074 -075 /** -076 * Gets the {@link org.apache.hadoop.hbase.client.TableDescriptor table descriptor} for this table. -077 * @throws java.io.IOException if a remote or network exception occurs. -078 */ -079 TableDescriptor getDescriptor() throws IOException; -080 -081 /** -082 * Test for the existence of columns in the table, as specified by the Get. -083 *

-084 * -085 * This will return true if the Get matches one or more keys, false if not. -086 *

-087 * -088 * This is a server-side call so it prevents any data from being transfered to -089 * the client. -090 * -091 * @param get the Get -092 * @return true if the specified Get matches one or more keys, false if not -093 * @throws IOException e -094 */ -095 boolean exists(Get get) throws IOException; -096 -097 /** -098 * Test for the existence of columns in the table, as specified by the Gets. -099 *

-100 * -101 * This will return an array of booleans. Each value will be true if the related Get matches -102 * one or more keys, false if not. -103 *

-104 * -105 * This is a server-side call so it prevents any data from being transferred to -106 * the client. -107 * -108 * @param gets the Gets -109 * @return Array of boolean. True if the specified Get matches one or more keys, false if not. -110 * @throws IOException e -111 */ -112 boolean[] existsAll(List gets) throws IOException; -113 -114 /** -115 * Method that does a batch call on Deletes, Gets, Puts, Increments, Appends, RowMutations. -116 * The ordering of execution of the actions is not defined. Meaning if you do a Put and a -117 * Get in the same {@link #batch} call, you will not necessarily be -118 * guaranteed that the Get returns what the Put had put. -119 * -120 * @param actions list of Get, Put, Delete, Increment, Append, RowMutations. -121 * @param results Empty Object[], same size as actions. Provides access to partial -122 *results, in case an exception is thrown. A null in the result array means that -123 *the call for that action failed, even after retries. The order of the objects -124 *in the results array corresponds to the order of actions in the request list. -125 * @throws IOException -126 * @since 0.90.0 -127 */ -128 void batch(final List actions, final Object[] results) throws IOException, -129InterruptedException; -130 -131 /** -132 * Same


[46/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/org/apache/hadoop/hbase/client/AsyncTableBase.html
--
diff --git a/apidocs/org/apache/hadoop/hbase/client/AsyncTableBase.html 
b/apidocs/org/apache/hadoop/hbase/client/AsyncTableBase.html
index 4b94f91..c564c00 100644
--- a/apidocs/org/apache/hadoop/hbase/client/AsyncTableBase.html
+++ b/apidocs/org/apache/hadoop/hbase/client/AsyncTableBase.html
@@ -355,10 +355,14 @@ public interface 
 
 getRpcTimeout
-long getRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
 title="class or interface in 
java.util.concurrent">TimeUnit unit)
+long getRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
 title="class or interface in 
java.util.concurrent">TimeUnit unit)
 Get timeout of each rpc request in this Table instance. It 
will be overridden by a more
  specific rpc timeout config such as readRpcTimeout or writeRpcTimeout.
 
+Parameters:
+unit - the unit of time the timeout to be represented in
+Returns:
+rpc timeout in the specified time unit
 See Also:
 getReadRpcTimeout(TimeUnit),
 
 getWriteRpcTimeout(TimeUnit)
@@ -371,8 +375,14 @@ public interface 
 
 getReadRpcTimeout
-long getReadRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
 title="class or interface in 
java.util.concurrent">TimeUnit unit)
+long getReadRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
 title="class or interface in 
java.util.concurrent">TimeUnit unit)
 Get timeout of each rpc read request in this Table 
instance.
+
+Parameters:
+unit - the unit of time the timeout to be represented in
+Returns:
+read rpc timeout in the specified time unit
+
 
 
 
@@ -381,8 +391,14 @@ public interface 
 
 getWriteRpcTimeout
-long getWriteRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
 title="class or interface in 
java.util.concurrent">TimeUnit unit)
+long getWriteRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
 title="class or interface in 
java.util.concurrent">TimeUnit unit)
 Get timeout of each rpc write request in this Table 
instance.
+
+Parameters:
+unit - the unit of time the timeout to be represented in
+Returns:
+write rpc timeout in the specified time unit
+
 
 
 
@@ -391,8 +407,14 @@ public interface 
 
 getOperationTimeout
-long getOperationTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
 title="class or interface in 
java.util.concurrent">TimeUnit unit)
+long getOperationTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
 title="class or interface in 
java.util.concurrent">TimeUnit unit)
 Get timeout of each operation in Table instance.
+
+Parameters:
+unit - the unit of time the timeout to be represented in
+Returns:
+operation rpc timeout in the specified time unit
+
 
 
 
@@ -401,9 +423,15 @@ public interface 
 
 getScanTimeout
-long getScanTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
 title="class or interface in 
java.util.concurrent">TimeUnit unit)
+long getScanTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
 title="class or interface in 
java.util.concurrent">TimeUnit unit)
 Get the timeout of a single operation in a scan. It works 
like operation timeout for other
  operations.
+
+Parameters:
+unit - the unit of time the timeout to be represented in
+Returns:
+scan rpc timeout in the specified time unit
+
 
 
 
@@ -412,7 +440,7 @@ public interface 
 
 exists
-default http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
 title="class or interface in java.util.concurrent">CompletableFutureBoolean> exists(Get get)
+default http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
 title="class or interface in java.util.concurrent">CompletableFutureBoolean> exists(Get get)
 Test for the existence of columns in the table, as 
specified by the Get.
  
  This will return true if the Get matches one or more keys, false if not.
@@ -431,7 +459,7 @@ public interface 
 
 get
-http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
 title="class or interface in java.util.concurrent">CompletableFuture get(Get get)
+http://docs.oracle.com/javase/8/docs/api/java/util/c

[11/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/devapidocs/org/apache/hadoop/hbase/coprocessor/MasterObserver.html
--
diff --git a/devapidocs/org/apache/hadoop/hbase/coprocessor/MasterObserver.html 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/MasterObserver.html
index b04e037..e58a62c 100644
--- a/devapidocs/org/apache/hadoop/hbase/coprocessor/MasterObserver.html
+++ b/devapidocs/org/apache/hadoop/hbase/coprocessor/MasterObserver.html
@@ -107,7 +107,7 @@ var activeTableTab = "activeTableTab";
 
 @InterfaceAudience.LimitedPrivate(value="Coprocesssor")
  @InterfaceStability.Evolving
-public interface MasterObserver
+public interface MasterObserver
 Defines coprocessor hooks for interacting with operations 
on the
  HMaster 
process.
  
@@ -351,8 +351,7 @@ public interface 
 default void
-postGetLocks(ObserverContext ctx,
-http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List lockedResources)
+postGetLocks(ObserverContext ctx)
 Called after a getLocks request has been processed.
 
 
@@ -365,8 +364,7 @@ public interface 
 default void
-postGetProcedures(ObserverContext ctx,
- http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
 title="class or interface in java.util">List> procList)
+postGetProcedures(ObserverContext ctx)
 Called after a getProcedures request has been 
processed.
 
 
@@ -429,9 +427,7 @@ public interface 
 default void
-postLockHeartbeat(ObserverContext ctx,
- LockProcedure proc,
- boolean keepAlive)
+postLockHeartbeat(ObserverContext ctx)
 Called after heartbeat to a lock.
 
 
@@ -530,11 +526,10 @@ public interface 
 default void
-postRequestLock(ObserverContext ctx,
+postRequestLock(ObserverContext ctx,
http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String namespace,
TableName tableName,
RegionInfo[] regionInfos,
-   LockType type,
http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String description)
 Called after new LockProcedure is queued.
 
@@ -657,8 +652,7 @@ public interface 
 default void
-preAbortProcedure(ObserverContext ctx,
- ProcedureExecutor procEnv,
+preAbortProcedure(ObserverContext ctx,
  long procId)
 Called before a abortProcedure request has been 
processed.
 
@@ -906,9 +900,9 @@ public interface 
 default void
-preLockHeartbeat(ObserverContext ctx,
-LockProcedure proc,
-boolean keepAlive)
+preLockHeartbeat(ObserverContext ctx,
+TableName tn,
+http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String description)
 Called before heartbeat to a lock.
 
 
@@ -1028,11 +1022,10 @@ public interface 
 default void
-preRequestLock(ObserverContext ctx,
+preRequestLock(ObserverContext ctx,
   http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String namespace,
   TableName tableName,
   RegionInfo[] regionInfos,
-  LockType type,
   http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String description)
 Called before new LockProcedure is queued.
 
@@ -1209,7 +1202,7 @@ public interface 
 
 preCreateTable
-default void preCreateTable(ObserverContext ctx,
+default void preCreateTable(ObserverContext ctx,
 TableDescriptor desc,
 RegionInfo[] regions)
  throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
@@ -1233,7 +1226,7 @@ public interface 
 
 postCreateTable
-default void postCreateTable(ObserverContext ctx,
+default void postCreateTable(ObserverContext ctx,
  TableDescriptor desc,
  RegionInfo[] regions)
   throws http://docs.oracle.com/

[40/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html
--
diff --git 
a/apidocs/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html 
b/apidocs/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html
index 47b2e30..a3890c3 100644
--- a/apidocs/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html
+++ b/apidocs/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html
@@ -18,8 +18,8 @@
 catch(err) {
 }
 //-->
-var methods = 
{"i0":9,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":9,"i8":10,"i9":10,"i10":10};
-var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
+var methods = 
{"i0":9,"i1":10,"i2":10,"i3":10,"i4":42,"i5":10,"i6":10,"i7":10,"i8":9,"i9":10,"i10":10,"i11":10};
+var tabs = {65535:["t0","All Methods"],1:["t1","Static 
Methods"],2:["t2","Instance Methods"],8:["t4","Concrete 
Methods"],32:["t6","Deprecated Methods"]};
 var altColor = "altColor";
 var rowColor = "rowColor";
 var tableTab = "tableTab";
@@ -200,7 +200,7 @@ extends org.apache.hadoop.hbase.filter.FilterBase
 
 Method Summary
 
-All Methods Static Methods Instance Methods Concrete Methods 
+All Methods Static Methods Instance Methods Concrete Methods Deprecated Methods 
 
 Modifier and Type
 Method and Description
@@ -215,48 +215,54 @@ extends org.apache.hadoop.hbase.filter.FilterBase
 
 
 Filter.ReturnCode
-filterColumn(Cell cell) 
+filterCell(Cell c)
+A way to filter based on the column family, column 
qualifier and/or the column value.
+
 
 
 Filter.ReturnCode
-filterKeyValue(Cell kv)
-A way to filter based on the column family, column 
qualifier and/or the column value.
-
+filterColumn(Cell cell) 
 
 
+Filter.ReturnCode
+filterKeyValue(Cell c)
+Deprecated. 
+
+
+
 boolean
 filterRowKey(Cell cell)
 Filters a row based on the row key.
 
 
-
+
 Cell
 getNextCellHint(Cell cell)
 Filters that are not sure which key must be next seeked to, 
can inherit
  this implementation that, by default, returns a null Cell.
 
 
-
+
 byte[][]
 getPrefix() 
 
-
+
 static MultipleColumnPrefixFilter
 parseFrom(byte[] pbBytes) 
 
-
+
 byte[]
 toByteArray()
 Return length 0 byte array for Filters that don't require 
special serialization
 
 
-
+
 http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
 toString()
 Return filter's info for debugging and logging 
purpose.
 
 
-
+
 protected http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
 title="class or interface in java.lang">String
 toString(int maxPrefixes) 
 
@@ -358,7 +364,7 @@ extends org.apache.hadoop.hbase.filter.FilterBase
  throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException
 Description copied from 
class: Filter
 Filters a row based on the row key. If this returns true, 
the entire row will be excluded. If
- false, each KeyValue in the row will be passed to Filter.filterKeyValue(Cell)
 below.
+ false, each KeyValue in the row will be passed to Filter.filterCell(Cell)
 below.
  If Filter.filterAllRemaining()
 returns true, then Filter.filterRowKey(Cell)
 should
  also return true.
 
@@ -382,7 +388,9 @@ extends org.apache.hadoop.hbase.filter.FilterBase
 
 
 filterKeyValue
-public Filter.ReturnCode filterKeyValue(Cell kv)
+http://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true";
 title="class or interface in java.lang">@Deprecated
+public Filter.ReturnCode filterKeyValue(Cell c)
+Deprecated. 
 Description copied from 
class: Filter
 A way to filter based on the column family, column 
qualifier and/or the column value. Return
  code is described below. This allows filters to filter only certain number of 
columns, then
@@ -395,14 +403,48 @@ extends org.apache.hadoop.hbase.filter.FilterBase
  If your filter returns ReturnCode.NEXT_ROW, it should return
  ReturnCode.NEXT_ROW until Filter.reset()
 is called just in case the caller calls
  for the next row.
- 
+
  Concrete implementers can signal a failure condition in their code by 
throwing an
  http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
 title="class or interface in java.io">IOException.
 
-Specified by:
+Overrides:
 filterKeyValue in
 class Filter
 Parameters:
-kv - the Cell in question
+c - the Cell in question
+Returns:
+code as described below, Filter.ReturnCode.INCLUDE by default
+See Also:
+Filter.ReturnCode
+
+
+
+
+
+
+
+
+filterCell
+public Filter.ReturnCode filterCell(Cell c)
+Description copied from 
class: Filter
+A way to filter based on the column family, column 
qualifier and/or the column value. Return
+ code is described below. This allows filters to filter only certain number of 
columns, t

[35/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/src-html/org/apache/hadoop/hbase/client/AsyncTableBase.html
--
diff --git 
a/apidocs/src-html/org/apache/hadoop/hbase/client/AsyncTableBase.html 
b/apidocs/src-html/org/apache/hadoop/hbase/client/AsyncTableBase.html
index f5f24c5..40ba9ea 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/client/AsyncTableBase.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/client/AsyncTableBase.html
@@ -71,345 +71,355 @@
 063   * specific rpc timeout config such as 
readRpcTimeout or writeRpcTimeout.
 064   * @see #getReadRpcTimeout(TimeUnit)
 065   * @see #getWriteRpcTimeout(TimeUnit)
-066   */
-067  long getRpcTimeout(TimeUnit unit);
-068
-069  /**
-070   * Get timeout of each rpc read request 
in this Table instance.
-071   */
-072  long getReadRpcTimeout(TimeUnit 
unit);
-073
-074  /**
-075   * Get timeout of each rpc write 
request in this Table instance.
-076   */
-077  long getWriteRpcTimeout(TimeUnit 
unit);
-078
-079  /**
-080   * Get timeout of each operation in 
Table instance.
-081   */
-082  long getOperationTimeout(TimeUnit 
unit);
-083
-084  /**
-085   * Get the timeout of a single 
operation in a scan. It works like operation timeout for other
-086   * operations.
-087   */
-088  long getScanTimeout(TimeUnit unit);
-089
-090  /**
-091   * Test for the existence of columns in 
the table, as specified by the Get.
-092   * 

-093 * This will return true if the Get matches one or more keys, false if not. -094 *

-095 * This is a server-side call so it prevents any data from being transfered to the client. -096 * @return true if the specified Get matches one or more keys, false if not. The return value will -097 * be wrapped by a {@link CompletableFuture}. -098 */ -099 default CompletableFuture exists(Get get) { -100return get(toCheckExistenceOnly(get)).thenApply(r -> r.getExists()); -101 } -102 -103 /** -104 * Extracts certain cells from a given row. -105 * @param get The object that specifies what data to fetch and from which row. -106 * @return The data coming from the specified row, if it exists. If the row specified doesn't -107 * exist, the {@link Result} instance returned won't contain any -108 * {@link org.apache.hadoop.hbase.KeyValue}, as indicated by {@link Result#isEmpty()}. The -109 * return value will be wrapped by a {@link CompletableFuture}. -110 */ -111 CompletableFuture get(Get get); +066 * @param unit the unit of time the timeout to be represented in +067 * @return rpc timeout in the specified time unit +068 */ +069 long getRpcTimeout(TimeUnit unit); +070 +071 /** +072 * Get timeout of each rpc read request in this Table instance. +073 * @param unit the unit of time the timeout to be represented in +074 * @return read rpc timeout in the specified time unit +075 */ +076 long getReadRpcTimeout(TimeUnit unit); +077 +078 /** +079 * Get timeout of each rpc write request in this Table instance. +080 * @param unit the unit of time the timeout to be represented in +081 * @return write rpc timeout in the specified time unit +082 */ +083 long getWriteRpcTimeout(TimeUnit unit); +084 +085 /** +086 * Get timeout of each operation in Table instance. +087 * @param unit the unit of time the timeout to be represented in +088 * @return operation rpc timeout in the specified time unit +089 */ +090 long getOperationTimeout(TimeUnit unit); +091 +092 /** +093 * Get the timeout of a single operation in a scan. It works like operation timeout for other +094 * operations. +095 * @param unit the unit of time the timeout to be represented in +096 * @return scan rpc timeout in the specified time unit +097 */ +098 long getScanTimeout(TimeUnit unit); +099 +100 /** +101 * Test for the existence of columns in the table, as specified by the Get. +102 *

+103 * This will return true if the Get matches one or more keys, false if not. +104 *

+105 * This is a server-side call so it prevents any data from being transfered to the client. +106 * @return true if the specified Get matches one or more keys, false if not. The return value will +107 * be wrapped by a {@link CompletableFuture}. +108 */ +109 default CompletableFuture exists(Get get) { +110return get(toCheckExistenceOnly(get)).thenApply(r -> r.getExists()); +111 } 112 113 /** -114 * Puts some data to the table. -115 * @param put The data to put. -116 * @return A {@link CompletableFuture} that always returns null when complete normally. -117 */ -118 CompletableFuture put(Put put); -119 -120 /** -121 * Deletes the specified cells/row. -122 * @param delete The object that specifies what to delete. -123 * @return A {@link CompletableFuture} that always returns null when complete normally


[24/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html
--
diff --git a/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html 
b/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html
index a38cce6..bf21363 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/RowFilter.html
@@ -84,83 +84,89 @@
 076this.filterOutRow = false;
 077  }
 078
-079  @Override
-080  public ReturnCode filterKeyValue(Cell 
v) {
-081if(this.filterOutRow) {
-082  return ReturnCode.NEXT_ROW;
-083}
-084return ReturnCode.INCLUDE;
-085  }
-086
-087  @Override
-088  public boolean filterRowKey(Cell 
firstRowCell) {
-089if (compareRow(getCompareOperator(), 
this.comparator, firstRowCell)) {
-090  this.filterOutRow = true;
-091}
-092return this.filterOutRow;
-093  }
-094
-095  @Override
-096  public boolean filterRow() {
-097return this.filterOutRow;
-098  }
-099
-100  public static Filter 
createFilterFromArguments(ArrayList filterArguments) {
-101@SuppressWarnings("rawtypes") // for 
arguments
-102ArrayList arguments = 
CompareFilter.extractArguments(filterArguments);
-103CompareOperator compareOp = 
(CompareOperator)arguments.get(0);
-104ByteArrayComparable comparator = 
(ByteArrayComparable)arguments.get(1);
-105return new RowFilter(compareOp, 
comparator);
-106  }
-107
-108 /**
-109  * @return The filter serialized using 
pb
-110  */
-111  public byte [] toByteArray() {
-112FilterProtos.RowFilter.Builder 
builder =
-113  
FilterProtos.RowFilter.newBuilder();
-114
builder.setCompareFilter(super.convert());
-115return 
builder.build().toByteArray();
-116  }
-117
-118  /**
-119   * @param pbBytes A pb serialized 
{@link RowFilter} instance
-120   * @return An instance of {@link 
RowFilter} made from bytes
-121   * @throws DeserializationException
-122   * @see #toByteArray
-123   */
-124  public static RowFilter parseFrom(final 
byte [] pbBytes)
-125  throws DeserializationException {
-126FilterProtos.RowFilter proto;
-127try {
-128  proto = 
FilterProtos.RowFilter.parseFrom(pbBytes);
-129} catch 
(InvalidProtocolBufferException e) {
-130  throw new 
DeserializationException(e);
-131}
-132final CompareOperator valueCompareOp 
=
-133  
CompareOperator.valueOf(proto.getCompareFilter().getCompareOp().name());
-134ByteArrayComparable valueComparator = 
null;
-135try {
-136  if 
(proto.getCompareFilter().hasComparator()) {
-137valueComparator = 
ProtobufUtil.toComparator(proto.getCompareFilter().getComparator());
-138  }
-139} catch (IOException ioe) {
-140  throw new 
DeserializationException(ioe);
-141}
-142return new 
RowFilter(valueCompareOp,valueComparator);
-143  }
-144
-145  /**
-146   * @return true if and only if the 
fields of the filter that are serialized
-147   * are equal to the corresponding 
fields in other.  Used for testing.
-148   */
-149  boolean areSerializedFieldsEqual(Filter 
o) {
-150if (o == this) return true;
-151if (!(o instanceof RowFilter)) return 
false;
-152
-153return 
super.areSerializedFieldsEqual(o);
-154  }
-155}
+079  @Deprecated
+080  @Override
+081  public ReturnCode filterKeyValue(final 
Cell c) {
+082return filterCell(c);
+083  }
+084
+085  @Override
+086  public ReturnCode filterCell(final Cell 
v) {
+087if(this.filterOutRow) {
+088  return ReturnCode.NEXT_ROW;
+089}
+090return ReturnCode.INCLUDE;
+091  }
+092
+093  @Override
+094  public boolean filterRowKey(Cell 
firstRowCell) {
+095if (compareRow(getCompareOperator(), 
this.comparator, firstRowCell)) {
+096  this.filterOutRow = true;
+097}
+098return this.filterOutRow;
+099  }
+100
+101  @Override
+102  public boolean filterRow() {
+103return this.filterOutRow;
+104  }
+105
+106  public static Filter 
createFilterFromArguments(ArrayList filterArguments) {
+107@SuppressWarnings("rawtypes") // for 
arguments
+108ArrayList arguments = 
CompareFilter.extractArguments(filterArguments);
+109CompareOperator compareOp = 
(CompareOperator)arguments.get(0);
+110ByteArrayComparable comparator = 
(ByteArrayComparable)arguments.get(1);
+111return new RowFilter(compareOp, 
comparator);
+112  }
+113
+114 /**
+115  * @return The filter serialized using 
pb
+116  */
+117  public byte [] toByteArray() {
+118FilterProtos.RowFilter.Builder 
builder =
+119  
FilterProtos.RowFilter.newBuilder();
+120
builder.setCompareFilter(super.convert());
+121return 
builder.build().toByteArray();
+122  }
+123
+124  /**
+125   * @param pbBytes A pb serialized 
{@link RowFilter} instance
+126   * @return An instance of {@link 
RowFilter} made from bytes
+127   * @throws Deserializat

[18/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/devapidocs/index-all.html
--
diff --git a/devapidocs/index-all.html b/devapidocs/index-all.html
index 6aee8ec..8a7db80 100644
--- a/devapidocs/index-all.html
+++ b/devapidocs/index-all.html
@@ -1112,7 +1112,7 @@
  
 add(Cell)
 - Method in class org.apache.hadoop.hbase.regionserver.querymatcher.ScanDeleteTracker
 
-Add the specified KeyValue to the list of deletes to check 
against for this row operation.
+Add the specified Cell to the list of deletes to check 
against for this row operation.
 
 add(String)
 - Method in class org.apache.hadoop.hbase.rest.client.Cluster
 
@@ -30004,6 +30004,66 @@
 
 filterBySubject(TableName)
 - Method in class org.apache.hadoop.hbase.quotas.TableQuotaSnapshotStore
  
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ColumnCountGetFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ColumnPaginationFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ColumnPrefixFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ColumnRangeFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.DependentColumnFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.FamilyFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.Filter
+
+A way to filter based on the column family, column 
qualifier and/or the column value.
+
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.FilterList
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.FilterListBase
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.FilterWrapper
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.FirstKeyOnlyFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.FirstKeyValueMatchingQualifiersFilter
+
+Deprecated.
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.FuzzyRowFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.InclusiveStopFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.KeyOnlyFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.MultipleColumnPrefixFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.MultiRowRangeFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.PageFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.PrefixFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.QualifierFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.RandomRowFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.RowFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.SingleColumnValueFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.SkipFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.TimestampsFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ValueFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.filter.WhileMatchFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.regionserver.MobReferenceOnlyFilter
+ 
 filterCell(Cell,
 Predicate) - Method in class 
org.apache.hadoop.hbase.replication.BulkLoadCellFilter
 
 Filters the bulk load cell using the supplied 
predicate.
@@ -30016,6 +30076,12 @@
 
 Applies the filter, possibly returning a different Cell 
instance.
 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.security.access.AccessControlFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.security.visibility.VisibilityController.DeleteVersionVisibilityExpressionFilter
+ 
+filterCell(Cell)
 - Method in class org.apache.hadoop.hbase.security.visibility.VisibilityLabelFilter
+ 
 filterCellByStore(WAL.Entry)
 - Method in class org.apache.hadoop.hbase.wal.WALSplitter.LogRecoveredEditsOutputSink
  
 filterCells(Result,
 Cell) - Static method in class org.apache.hadoop.hbase.client.ConnectionUtils
@@ -30074,71 +30140,114 @@
  reports received from RegionServers yet.
 
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ColumnCountGetFilter
- 
+
+Deprecated.
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ColumnPaginationFilter
- 
+
+Deprecated.
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ColumnPrefixFilter
- 
+
+Deprecated.
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.ColumnRangeFilter
- 
+
+Deprecated.
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.DependentColumnFilter
- 
+
+Deprecated.
+
 filterKeyValue(Cell)
 - Method in class org.apache.hadoop.hbase.filter.FamilyFilter
- 
+
+Depreca

[50/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apache_hbase_reference_guide.pdf
--
diff --git a/apache_hbase_reference_guide.pdf b/apache_hbase_reference_guide.pdf
index 410b1af..61a220e 100644
--- a/apache_hbase_reference_guide.pdf
+++ b/apache_hbase_reference_guide.pdf
@@ -5,16 +5,16 @@
 /Author (Apache HBase Team)
 /Creator (Asciidoctor PDF 1.5.0.alpha.15, based on Prawn 2.2.2)
 /Producer (Apache HBase Team)
-/ModDate (D:20171025144654+00'00')
-/CreationDate (D:20171025144654+00'00')
+/ModDate (D:20171027144649+00'00')
+/CreationDate (D:20171027144649+00'00')
 >>
 endobj
 2 0 obj
 << /Type /Catalog
 /Pages 3 0 R
 /Names 26 0 R
-/Outlines 4321 0 R
-/PageLabels 4529 0 R
+/Outlines 4325 0 R
+/PageLabels 4533 0 R
 /PageMode /UseOutlines
 /OpenAction [7 0 R /FitH 842.89]
 /ViewerPreferences << /DisplayDocTitle true
@@ -23,8 +23,8 @@ endobj
 endobj
 3 0 obj
 << /Type /Pages
-/Count 666
-/Kids [7 0 R 12 0 R 14 0 R 16 0 R 18 0 R 20 0 R 22 0 R 24 0 R 44 0 R 47 0 R 50 
0 R 54 0 R 63 0 R 66 0 R 69 0 R 71 0 R 76 0 R 80 0 R 83 0 R 89 0 R 91 0 R 94 0 
R 96 0 R 103 0 R 109 0 R 114 0 R 116 0 R 130 0 R 133 0 R 142 0 R 151 0 R 161 0 
R 170 0 R 181 0 R 185 0 R 187 0 R 191 0 R 202 0 R 209 0 R 218 0 R 226 0 R 231 0 
R 240 0 R 248 0 R 259 0 R 271 0 R 278 0 R 287 0 R 296 0 R 304 0 R 310 0 R 319 0 
R 325 0 R 332 0 R 339 0 R 348 0 R 357 0 R 366 0 R 379 0 R 387 0 R 394 0 R 402 0 
R 409 0 R 417 0 R 427 0 R 434 0 R 442 0 R 451 0 R 464 0 R 470 0 R 477 0 R 485 0 
R 493 0 R 502 0 R 510 0 R 515 0 R 520 0 R 525 0 R 541 0 R 551 0 R 556 0 R 570 0 
R 576 0 R 581 0 R 583 0 R 585 0 R 590 0 R 598 0 R 604 0 R 609 0 R 614 0 R 625 0 
R 636 0 R 641 0 R 660 0 R 675 0 R 686 0 R 688 0 R 690 0 R 698 0 R 710 0 R 719 0 
R 729 0 R 735 0 R 738 0 R 742 0 R 747 0 R 750 0 R 753 0 R 755 0 R 758 0 R 762 0 
R 764 0 R 769 0 R 773 0 R 778 0 R 783 0 R 786 0 R 792 0 R 794 0 R 798 0 R 807 0 
R 809 0 R 812 0 R 815 0 R 818 0 R 821 0 
 R 835 0 R 842 0 R 851 0 R 862 0 R 868 0 R 880 0 R 884 0 R 887 0 R 891 0 R 894 
0 R 899 0 R 908 0 R 916 0 R 920 0 R 924 0 R 929 0 R 933 0 R 935 0 R 950 0 R 961 
0 R 967 0 R 973 0 R 976 0 R 984 0 R 992 0 R 996 0 R 1002 0 R 1007 0 R 1009 0 R 
1011 0 R 1013 0 R 1024 0 R 1032 0 R 1036 0 R 1043 0 R 1051 0 R 1059 0 R 1063 0 
R 1069 0 R 1074 0 R 1082 0 R 1087 0 R 1092 0 R 1094 0 R 1100 0 R 1106 0 R 1108 
0 R 1115 0 R 1125 0 R 1129 0 R 1131 0 R 1133 0 R 1137 0 R 1140 0 R 1145 0 R 
1148 0 R 1160 0 R 1164 0 R 1170 0 R 1177 0 R 1182 0 R 1186 0 R 1190 0 R 1192 0 
R 1195 0 R 1198 0 R 1201 0 R 1205 0 R 1209 0 R 1213 0 R 1218 0 R 1222 0 R 1225 
0 R 1227 0 R 1239 0 R 1242 0 R 1250 0 R 1259 0 R 1265 0 R 1269 0 R 1271 0 R 
1281 0 R 1284 0 R 1290 0 R 1298 0 R 1301 0 R 1308 0 R 1317 0 R 1319 0 R 1321 0 
R 1331 0 R 1333 0 R 1335 0 R 1338 0 R 1340 0 R 1342 0 R 1344 0 R 1346 0 R 1349 
0 R 1353 0 R 1358 0 R 1360 0 R 1362 0 R 1364 0 R 1369 0 R 1376 0 R 1381 0 R 
1384 0 R 1386 0 R 1389 0 R 1393 0 R 1395 0 R 1398 0 R 1400
  0 R 1402 0 R 1405 0 R 1410 0 R 1415 0 R 1424 0 R 1438 0 R 1452 0 R 1455 0 R 
1459 0 R 1473 0 R 1482 0 R 1496 0 R 1502 0 R 1510 0 R 1525 0 R 1539 0 R 1551 0 
R 1556 0 R 1562 0 R 1573 0 R 1579 0 R 1585 0 R 1593 0 R 1596 0 R 1605 0 R 1612 
0 R 1616 0 R 1629 0 R 1631 0 R 1637 0 R 1643 0 R 1647 0 R 1655 0 R 1664 0 R 
1668 0 R 1670 0 R 1672 0 R 1685 0 R 1691 0 R 1699 0 R 1706 0 R 1720 0 R 1725 0 
R 1734 0 R 1742 0 R 1748 0 R 1755 0 R 1759 0 R 1762 0 R 1764 0 R 1770 0 R 1776 
0 R 1782 0 R 1786 0 R 1794 0 R 1799 0 R 1805 0 R 1810 0 R 1812 0 R 1820 0 R 
1828 0 R 1834 0 R 1839 0 R 1843 0 R 1846 0 R 1851 0 R 1856 0 R 1863 0 R 1865 0 
R 1867 0 R 1870 0 R 1878 0 R 1881 0 R 1888 0 R 1898 0 R 1901 0 R 1906 0 R 1908 
0 R 1913 0 R 1916 0 R 1918 0 R 1923 0 R 1933 0 R 1935 0 R 1937 0 R 1939 0 R 
1941 0 R 1944 0 R 1946 0 R 1948 0 R 1951 0 R 1953 0 R 1955 0 R 1959 0 R 1963 0 
R 1972 0 R 1974 0 R 1976 0 R 1982 0 R 1984 0 R 1989 0 R 1991 0 R 1993 0 R 2000 
0 R 2005 0 R 2009 0 R 2013 0 R 2017 0 R 2019 0 R 2021 0 R 20
 25 0 R 2028 0 R 2030 0 R 2032 0 R 2036 0 R 2038 0 R 2041 0 R 2043 0 R 2045 0 R 
2047 0 R 2054 0 R 2057 0 R 2062 0 R 2064 0 R 2066 0 R 2068 0 R 2070 0 R 2078 0 
R 2089 0 R 2103 0 R 2114 0 R 2118 0 R 2123 0 R 2127 0 R 2130 0 R 2135 0 R 2141 
0 R 2143 0 R 2147 0 R 2149 0 R 2151 0 R 2153 0 R 2157 0 R 2159 0 R 2172 0 R 
2175 0 R 2183 0 R 2189 0 R 2201 0 R 2215 0 R 2229 0 R 2246 0 R 2250 0 R 2252 0 
R 2256 0 R 2274 0 R 2280 0 R 2292 0 R 2296 0 R 2300 0 R 2309 0 R 2319 0 R 2324 
0 R 2335 0 R 2348 0 R 2367 0 R 2376 0 R 2379 0 R 2388 0 R 2406 0 R 2413 0 R 
2416 0 R 2421 0 R 2425 0 R 2428 0 R 2437 0 R 2446 0 R 2449 0 R 2451 0 R 2455 0 
R 2470 0 R 2478 0 R 2483 0 R 2487 0 R 2491 0 R 2493 0 R 2495 0 R 2497 0 R 2502 
0 R 2515 0 R 2525 0 R 2534 0 R 2543 0 R 2549 0 R 2560 0 R 2567 0 R 2573 0 R 
2575 0 R 2585 0 R 2593 0 R 2603 0 R 2607 0 R 2618 0 R 2622 0 R 2632 0 R 2640 0 
R 2648 0 R 2654 0 R 2658 0 R 2662 0 R 2666 0 R 2668 0 R 2674 0 R 2678 0 R 2682 
0 R 2688 0 R 2694 0 R 2697 0 R 27

[08/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleMasterObserverWithMetrics.html
--
diff --git 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleMasterObserverWithMetrics.html
 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleMasterObserverWithMetrics.html
index f80f327..2151827 100644
--- 
a/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleMasterObserverWithMetrics.html
+++ 
b/devapidocs/org/apache/hadoop/hbase/coprocessor/example/ExampleMasterObserverWithMetrics.html
@@ -287,7 +287,7 @@ implements MasterObserver
-postAbortProcedure,
 postAddReplicationPeer,
 postAddRSGroup,
 postAssign,
 postBalance, postBalanceRSGroup,
 postBalanceSwitch,
 postClearDeadServers,
 postCloneSnapshot,
 postCompletedCreateTableAction,
 postCompletedDeleteTableAction,
 postCompletedDisableTableAction,
 postCompletedEnableTableAction,
 postCompletedMergeRegionsAction,
 postCompletedModifyTableAction,
 postCompletedSplitRegionAction,
 postCompletedTruncateTab
 leAction, postCreateNamespace,
 postDecommissionRegionServers,
 postDeleteNamespace,
 postDeleteSnapshot,
 postDeleteTable,
 postDisableReplicationPeer,
 postDisableTable,
 postEnableReplicationPeer,
 postEnableTable,
 postGetLocks,
 postGetNamespaceDescriptor,
 postGetProcedures,
 postGetReplicationPeerConfig,
 postG
 etTableDescriptors, postGetTableNames,
 postListDeadServers,
 postListDecommissionedRegionServers,
 postListNamespaceDescriptors,
 postListReplicationPeers, postListSnapshot,
 postLockHeartbeat,
 postMergeRegions,
 po
 stMergeRegionsCommitAction, postModifyNamespace,
 postModifyTable,
 postMove,
 postMoveServers,
 postMoveServersAndTables,
 postMoveTables,
 postRecommissionRegionServer,
 postRegionOffline,
 postRemoveReplicationPeer,
 postRemoveRSGroup,
 postRequestLock,
 postRestoreSnapshot,
 postRollBackMergeRegionsAction,
 postRollBackSplitRegionAction,
 postSetNamespaceQuota,
 postSetSplitOrMergeEnabled,
 postSetTableQuota,
 postSetUserQuota,
 postSetUserQuota,
 postSetUserQuota,
 postSnapshot,
 postStartMaster,
 postTableFlush,
 postTruncateTable,
 postUn
 assign, postUpdateReplicationPeerConfig,
 preAbortProcedure,
 preAddReplicationPeer,
 preAddRSGroup,
 preAssign,
 preBalance,
 preBalanceRSGroup,
 preBalanceSwitch,
 preClearDeadServers,
 preCloneSnapshot,
 preCreateNamespace,
 preCreateTableAction,
 preDecommissionRegionServers,
 preDeleteNamespace,
 preDeleteSnapshot,
 preDeleteTable,
 preDeleteTableAction,
 preDisableReplicationPeer,
 preDisableTableAction,
 preEnableReplicationPeer,
 preEnableTable,
 preEnableTableAction,
 preGetLocks,
 preGetNamespaceDescriptor,
 preGetProcedures,
 preGetReplicationPeerConfig,
 preGetTableDescriptors,
 preGetTableNames, preListDeadServers,
 preListDecommissionedRegionServers,
 preListNamespaceDescriptors,
 preListReplicationPeers,
 preListSnapshot,
 preLockHeartbeat,
 preMasterInitialization,
 preMergeRegions,
 preMergeRegionsAction,
 preMergeRegionsCommitAction,
 preModifyNamespace,
 preModifyTable,
 preModifyTableAction,
 preMove,
 preMoveServers,
 preMoveServersAndTables,
 preMoveTables,
 preRecommissionRegionServer,
 preRegionOffline,
 preRemoveReplicationPeer,
 preRemoveRSGroup,
 preRequestLock, preRestoreSnapshot,
 preSetNamespaceQuota,
 preSetSplitOrMergeEnabled,
 preSetTableQuota,
 preSetUserQuota,
 preSetUserQuota,
 preSetUserQuota,
 preShutdown, preSnapshot,
 preSplitRegion,
 preSplitRegionAction,
 preSplitRegionAfterMETAAction,
 preSplitRegionBeforeMETAAction,
 preStopMaster,
 preTableFlush,
 preTruncateTable,
 preTruncateTableActi
 on, preUnassign,
 preUpdateReplicationPeerConfig
+postAbortProcedure,
 postAddReplicationPeer,
 postAddRSGroup,
 postAssign,
 postBalance, postBalanceRSGroup,
 postBalanceSwitch,
 postClearDeadServers,
 postCloneSnapshot,
 postCompletedCreateTableAction,
 postCompletedDeleteTableAction,
 postCompletedDisableTableAction,
 postCompletedEnableTableAction,
 postCompletedMergeRegionsAction,
 postCompletedModifyTableAction,
 postCompletedSplitRegionAction,
 postCompletedTruncateTab
 leAction, postCreateNamespace,
 postDecommissionRegionServers,
 postDeleteNamespace,
 postDeleteSnapshot,
 postDeleteTable,
 postDisableReplicationPeer,
 postDisableTable,
 postEnableReplicationPeer,
 postEnableTable,
 postGetLocks,
 postGetNamespaceDescriptor,
 postGetProcedures,
 postGetReplicationPeerConfig,
 postGetTableDescriptors,
 postGetTableNames,
 postListDeadServers,
 postListDecommissionedRegionServers,
 postListNamespaceDescriptors,
 pos
 tListReplicationPeers, postListSnapshot,
 postLockHeartbeat,
 postMergeRegions,
 postMergeRegionsCommitAction,
 postModifyNamespace,
 postModifyTable,
 postMove,
 postMoveServers,
 postMoveServersAndTables,
 postMoveTables,
 postRecommissionRegionServer,
 postRegionOffline,
 postR
 emoveReplicationPeer, postRemoveRSGroup,
 postRequestLock,
 postRestoreSnapshot,
 postRo

[31/51] [partial] hbase-site git commit: Published site at .

2017-10-27 Thread git-site-role
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/src-html/org/apache/hadoop/hbase/filter/Filter.html
--
diff --git a/apidocs/src-html/org/apache/hadoop/hbase/filter/Filter.html 
b/apidocs/src-html/org/apache/hadoop/hbase/filter/Filter.html
index 35d409c..22238bc 100644
--- a/apidocs/src-html/org/apache/hadoop/hbase/filter/Filter.html
+++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/Filter.html
@@ -42,7 +42,7 @@
 034 *   
  • {@link #reset()} : reset the filter state before filtering a new row.
  • 035 *
  • {@link #filterAllRemaining()}: true means row scan is over; false means keep going.
  • 036 *
  • {@link #filterRowKey(Cell)}: true means drop this row; false means include.
  • -037 *
  • {@link #filterKeyValue(Cell)}: decides whether to include or exclude this Cell. +037 *
  • {@link #filterCell(Cell)}: decides whether to include or exclude this Cell. 038 *See {@link ReturnCode}.
  • 039 *
  • {@link #transformCell(Cell)}: if the Cell is included, let the filter transform the 040 *Cell.
  • @@ -74,7 +74,7 @@ 066 067 /** 068 * Filters a row based on the row key. If this returns true, the entire row will be excluded. If -069 * false, each KeyValue in the row will be passed to {@link #filterKeyValue(Cell)} below. +069 * false, each KeyValue in the row will be passed to {@link #filterCell(Cell)} below. 070 * 071 * Concrete implementers can signal a failure condition in their code by throwing an 072 * {@link IOException}. @@ -92,7 +92,7 @@ 084 085 /** 086 * Filters a row based on the row key. If this returns true, the entire row will be excluded. If -087 * false, each KeyValue in the row will be passed to {@link #filterKeyValue(Cell)} below. +087 * false, each KeyValue in the row will be passed to {@link #filterCell(Cell)} below. 088 * If {@link #filterAllRemaining()} returns true, then {@link #filterRowKey(Cell)} should 089 * also return true. 090 * @@ -128,184 +128,214 @@ 120 * If your filter returns ReturnCode.NEXT_ROW, it should return 121 * ReturnCode.NEXT_ROW until {@link #reset()} is called just in case the caller calls 122 * for the next row. -123 * +123 * 124 * Concrete implementers can signal a failure condition in their code by throwing an 125 * {@link IOException}. 126 * -127 * @param v the Cell in question -128 * @return code as described below +127 * @param c the Cell in question +128 * @return code as described below, Filter.ReturnCode.INCLUDE by default 129 * @throws IOException in case an I/O or an filter specific failure needs to be signaled. 130 * @see Filter.ReturnCode -131 */ -132 abstract public ReturnCode filterKeyValue(final Cell v) throws IOException; -133 -134 /** -135 * Give the filter a chance to transform the passed KeyValue. If the Cell is changed a new -136 * Cell object must be returned. -137 * -138 * @see org.apache.hadoop.hbase.KeyValue#shallowCopy() -139 * The transformed KeyValue is what is eventually returned to the client. Most filters will -140 * return the passed KeyValue unchanged. -141 * @see org.apache.hadoop.hbase.filter.KeyOnlyFilter#transformCell(Cell) for an example of a -142 * transformation. -143 * -144 * Concrete implementers can signal a failure condition in their code by throwing an -145 * {@link IOException}. -146 * -147 * @param v the KeyValue in question -148 * @return the changed KeyValue -149 * @throws IOException in case an I/O or an filter specific failure needs to be signaled. -150 */ -151 abstract public Cell transformCell(final Cell v) throws IOException; -152 -153 /** -154 * Return codes for filterValue(). -155 */ -156 @InterfaceAudience.Public -157 public enum ReturnCode { -158/** -159 * Include the Cell -160 */ -161INCLUDE, -162/** -163 * Include the Cell and seek to the next column skipping older versions. -164 */ -165INCLUDE_AND_NEXT_COL, -166/** -167 * Skip this Cell -168 */ -169SKIP, -170/** -171 * Skip this column. Go to the next column in this row. -172 */ -173NEXT_COL, -174/** -175 * Seek to next row in current family. It may still pass a cell whose family is different but -176 * row is the same as previous cell to {@link #filterKeyValue(Cell)} , even if we get a NEXT_ROW -177 * returned for previous cell. For more details see HBASE-18368.
    -178 * Once reset() method was invoked, then we switch to the next row for all family, and you can -179 * catch the event by invoking CellUtils.matchingRows(previousCell, currentCell).
    -180 * Note that filterRow() will still be called.
    -181 */ -182NEXT_ROW, -183/** -184 * Seek to next key which is given as hint by the filter. -1

    [44/51] [partial] hbase-site git commit: Published site at .

    2017-10-27 Thread git-site-role
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.html
    --
    diff --git a/apidocs/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.html 
    b/apidocs/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.html
    index f5c3f03..b556dab 100644
    --- a/apidocs/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.html
    +++ b/apidocs/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.html
    @@ -18,8 +18,8 @@
     catch(err) {
     }
     //-->
    -var methods = 
    {"i0":9,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":9,"i8":10,"i9":10,"i10":10};
    -var tabs = {65535:["t0","All Methods"],1:["t1","Static 
    Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
    +var methods = 
    {"i0":9,"i1":10,"i2":42,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":9,"i9":10,"i10":10,"i11":10};
    +var tabs = {65535:["t0","All Methods"],1:["t1","Static 
    Methods"],2:["t2","Instance Methods"],8:["t4","Concrete 
    Methods"],32:["t6","Deprecated Methods"]};
     var altColor = "altColor";
     var rowColor = "rowColor";
     var tableTab = "tableTab";
    @@ -195,7 +195,7 @@ extends org.apache.hadoop.hbase.filter.FilterBase
     
     Method Summary
     
    -All Methods Static Methods Instance Methods Concrete Methods 
    +All Methods Static Methods Instance Methods Concrete Methods Deprecated Methods 
     
     Modifier and Type
     Method and Description
    @@ -206,53 +206,59 @@ extends org.apache.hadoop.hbase.filter.FilterBase
     
     
     Filter.ReturnCode
    -filterKeyValue(Cell v)
    +filterCell(Cell c)
     A way to filter based on the column family, column 
    qualifier and/or the column value.
     
     
     
    +Filter.ReturnCode
    +filterKeyValue(Cell c)
    +Deprecated. 
    +
    +
    +
     boolean
     filterRowKey(Cell cell)
     Filters a row based on the row key.
     
     
    -
    +
     byte[]
     getColumnOffset() 
     
    -
    +
     int
     getLimit() 
     
    -
    +
     Cell
     getNextCellHint(Cell cell)
     Filters that are not sure which key must be next seeked to, 
    can inherit
      this implementation that, by default, returns a null Cell.
     
     
    -
    +
     int
     getOffset() 
     
    -
    +
     static ColumnPaginationFilter
     parseFrom(byte[] pbBytes) 
     
    -
    +
     void
     reset()
     Filters that are purely stateless and do nothing in their 
    reset() methods can inherit
      this null/empty implementation.
     
     
    -
    +
     byte[]
     toByteArray()
     Return length 0 byte array for Filters that don't require 
    special serialization
     
     
    -
    +
     http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String
     toString()
     Return filter's info for debugging and logging 
    purpose.
    @@ -390,7 +396,7 @@ extends org.apache.hadoop.hbase.filter.FilterBase
      throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
     title="class or interface in java.io">IOException
     Description copied from 
    class: Filter
     Filters a row based on the row key. If this returns true, 
    the entire row will be excluded. If
    - false, each KeyValue in the row will be passed to Filter.filterKeyValue(Cell)
     below.
    + false, each KeyValue in the row will be passed to Filter.filterCell(Cell)
     below.
      If Filter.filterAllRemaining()
     returns true, then Filter.filterRowKey(Cell)
     should
      also return true.
     
    @@ -414,7 +420,9 @@ extends org.apache.hadoop.hbase.filter.FilterBase
     
     
     filterKeyValue
    -public Filter.ReturnCode filterKeyValue(Cell v)
    +http://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true";
     title="class or interface in java.lang">@Deprecated
    +public Filter.ReturnCode filterKeyValue(Cell c)
    +Deprecated. 
     Description copied from 
    class: Filter
     A way to filter based on the column family, column 
    qualifier and/or the column value. Return
      code is described below. This allows filters to filter only certain number of 
    columns, then
    @@ -427,14 +435,48 @@ extends org.apache.hadoop.hbase.filter.FilterBase
      If your filter returns ReturnCode.NEXT_ROW, it should return
      ReturnCode.NEXT_ROW until Filter.reset()
     is called just in case the caller calls
      for the next row.
    - 
    +
      Concrete implementers can signal a failure condition in their code by 
    throwing an
      http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
     title="class or interface in java.io">IOException.
     
    -Specified by:
    +Overrides:
     filterKeyValue in
     class Filter
     Parameters:
    -v - the Cell in question
    +c - the Cell in question
    +Returns:
    +code as described below, Filter.ReturnCode.INCLUDE by default
    +See Also:
    +Filter.ReturnCode
    +
    +
    +
    +
    +
    +
    +
    +
    +filterCell
    +public Filter.ReturnCode filterCell(Cell c)
    +Description copied from 
    class: Filter
    +A way to filter based on the column family, column 
    qualifier and/or the column value. Return
    + code is described below. This allows filters to filter only certain number of 
    columns, then
    + terminate without matching ever column.
    +
    + If filterRowKey returns true, filterCell needs to be consistent with it.
    +
    + filterCell can assume that filterRowKey has alr

    [19/51] [partial] hbase-site git commit: Published site at .

    2017-10-27 Thread git-site-role
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/checkstyle.rss
    --
    diff --git a/checkstyle.rss b/checkstyle.rss
    index d0f9173..158d6d9 100644
    --- a/checkstyle.rss
    +++ b/checkstyle.rss
    @@ -26,7 +26,7 @@ under the License.
     ©2007 - 2017 The Apache Software Foundation
     
       File: 2060,
    - Errors: 13620,
    + Errors: 13637,
      Warnings: 0,
      Infos: 0
       
    @@ -377,7 +377,7 @@ under the License.
       0
     
     
    -  18
    +  17
     
       
       
    @@ -517,7 +517,7 @@ under the License.
       0
     
     
    -  8
    +  9
     
       
       
    @@ -1007,7 +1007,7 @@ under the License.
       0
     
     
    -  54
    +  62
     
       
       
    @@ -5781,7 +5781,7 @@ under the License.
       0
     
     
    -  11
    +  12
     
       
       
    @@ -5949,7 +5949,7 @@ under the License.
       0
     
     
    -  10
    +  11
     
       
       
    @@ -7531,7 +7531,7 @@ under the License.
       0
     
     
    -  7
    +  8
     
       
       
    @@ -9519,7 +9519,7 @@ under the License.
       0
     
     
    -  8
    +  9
     
       
       
    @@ -9659,7 +9659,7 @@ under the License.
       0
     
     
    -  18
    +  19
     
       
       
    @@ -12865,7 +12865,7 @@ under the License.
       0
     
     
    -  7
    +  8
     
       
       
    @@ -14139,7 +14139,7 @@ under the License.
       0
     
     
    -  21
    +  22
     
       
       
    @@ -14489,7 +14489,7 @@ under the License.
       0
     
     
    -  22
    +  20
     
       
       
    @@ -17205,7 +17205,7 @@ under the License.
       0
     
     
    -  110
    +  113
     
       
       
    @@ -17233,7 +17233,7 @@ under the License.
       0
     
     
    -  6
    +  7
     
       
       
    @@ -20033,7 +20033,7 @@ under the License.
       0
     
     
    -  19
    +  17
     
       
       
    @@ -25115,7 +25115,7 @@ under the License.
       0
     
     
    -  30
    +  31
     
       
       
    @@ -25913,7 +25913,7 @@ under the License.
       0
     
     
    -  20
    +  21
     
       
       
    @@ -26165,7 +26165,7 @@ under the License.
       0
     
     
    -  222
    +  223
     
       
       
    @@ -27299,7 +27299,7 @@ under the License.
       0
     
     
    -  64
    +  62
     
       
       
    @@ -28657,7 +28657,7 @@ under the License.
       0
     
     
    -  8
    +  9
     
       
       
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/coc.html
    --
    diff --git a/coc.html b/coc.html
    index bbb76dd..01eeb8e 100644
    --- a/coc.html
    +++ b/coc.html
    @@ -7,7 +7,7 @@
       
     
     
    -
    +
     
     Apache HBase – 
       Code of Conduct Policy
    @@ -380,7 +380,7 @@ email to mailto:priv...@hbase.apache.org";>the priv
     https://www.apache.org/";>The Apache Software 
    Foundation.
     All rights reserved.  
     
    -  Last Published: 
    2017-10-25
    +  

    [20/51] [partial] hbase-site git commit: Published site at .

    2017-10-27 Thread git-site-role
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/checkstyle-aggregate.html
    --
    diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html
    index ddc9ff4..b298db3 100644
    --- a/checkstyle-aggregate.html
    +++ b/checkstyle-aggregate.html
    @@ -7,7 +7,7 @@
       
     
     
    -
    +
     
     Apache HBase – Checkstyle Results
     
    @@ -289,7 +289,7 @@
     2060
     0
     0
    -13620
    +13637
     
     Files
     
    @@ -1147,7 +1147,7 @@
     org/apache/hadoop/hbase/client/HTable.java
     0
     0
    -64
    +62
     
     org/apache/hadoop/hbase/client/HTableMultiplexer.java
     0
    @@ -1472,7 +1472,7 @@
     org/apache/hadoop/hbase/client/Table.java
     0
     0
    -19
    +17
     
     org/apache/hadoop/hbase/client/TableDescriptor.java
     0
    @@ -1932,7 +1932,7 @@
     org/apache/hadoop/hbase/coprocessor/MasterObserver.java
     0
     0
    -22
    +20
     
     org/apache/hadoop/hbase/coprocessor/MetricsCoprocessor.java
     0
    @@ -2152,7 +2152,7 @@
     org/apache/hadoop/hbase/filter/ColumnRangeFilter.java
     0
     0
    -20
    +21
     
     org/apache/hadoop/hbase/filter/CompareFilter.java
     0
    @@ -2162,12 +2162,12 @@
     org/apache/hadoop/hbase/filter/DependentColumnFilter.java
     0
     0
    -11
    +12
     
     org/apache/hadoop/hbase/filter/FamilyFilter.java
     0
     0
    -10
    +11
     
     org/apache/hadoop/hbase/filter/Filter.java
     0
    @@ -2182,7 +2182,7 @@
     org/apache/hadoop/hbase/filter/FilterList.java
     0
     0
    -7
    +8
     
     org/apache/hadoop/hbase/filter/FilterListBase.java
     0
    @@ -2217,7 +2217,7 @@
     org/apache/hadoop/hbase/filter/FuzzyRowFilter.java
     0
     0
    -18
    +19
     
     org/apache/hadoop/hbase/filter/InclusiveStopFilter.java
     0
    @@ -2237,7 +2237,7 @@
     org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
     0
     0
    -8
    +9
     
     org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java
     0
    @@ -2252,7 +2252,7 @@
     org/apache/hadoop/hbase/filter/PageFilter.java
     0
     0
    -8
    +9
     
     org/apache/hadoop/hbase/filter/ParseConstants.java
     0
    @@ -2272,7 +2272,7 @@
     org/apache/hadoop/hbase/filter/QualifierFilter.java
     0
     0
    -7
    +8
     
     org/apache/hadoop/hbase/filter/RandomRowFilter.java
     0
    @@ -2287,7 +2287,7 @@
     org/apache/hadoop/hbase/filter/RowFilter.java
     0
     0
    -6
    +7
     
     org/apache/hadoop/hbase/filter/SingleColumnValueExcludeFilter.java
     0
    @@ -2297,7 +2297,7 @@
     org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java
     0
     0
    -21
    +22
     
     org/apache/hadoop/hbase/filter/SkipFilter.java
     0
    @@ -2317,7 +2317,7 @@
     org/apache/hadoop/hbase/filter/ValueFilter.java
     0
     0
    -8
    +9
     
     org/apache/hadoop/hbase/filter/WhileMatchFilter.java
     0
    @@ -3637,7 +3637,7 @@
     org/apache/hadoop/hbase/master/MasterCoprocessorHost.java
     0
     0
    -18
    +17
     
     org/apache/hadoop/hbase/master/MasterDumpServlet.java
     0
    @@ -5042,7 +5042,7 @@
     org/apache/hadoop/hbase/regionserver/HRegion.java
     0
     0
    -222
    +223
     
     org/apache/hadoop/hbase/regionserver/HRegionFileSystem.java
     0
    @@ -5062,7 +5062,7 @@
     org/apache/hadoop/hbase/regionserver/HStore.java
     0
     0
    -54
    +62
     
     org/apache/hadoop/hbase/regionserver/HStoreFile.java
     0
    @@ -5332,7 +5332,7 @@
     org/apache/hadoop/hbase/regionserver/Region.java
     0
     0
    -30
    +31
     
     org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java
     0
    @@ -6407,7 +6407,7 @@
     org/apache/hadoop/hbase/rest/client/RemoteHTable.java
     0
     0
    -110
    +113
     
     org/apache/hadoop/hbase/rest/client/Response.java
     0
    @@ -8204,7 +8204,7 @@
     
     annotation
     http://checkstyle.sourceforge.net/config_annotation.html#MissingDeprecated";>MissingDeprecated
    -77
    +108
      Error
     
     blocks
    @@ -8214,12 +8214,12 @@
     
     
     http://checkstyle.sourceforge.net/config_blocks.html#LeftCurly";>LeftCurly
    -231
    +230
      Error
     
     
     http://checkstyle.sourceforge.net/config_blocks.html#NeedBraces";>NeedBraces
    -1694
    +1695
      Error
     
     coding
    @@ -8293,7 +8293,7 @@
     http://checkstyle.sourceforge.net/config_imports.html#UnusedImports";>UnusedImports
     
     processJavadoc: "true"
    -121
    +118
      Error
     
     indentation
    @@ -8304,19 +8304,19 @@
     caseIndent: "2"
     basicOffset: "2"
     lineWrappingIndentation: "2"
    -3711
    +3718
      Error
     
     javadoc
     http://checkstyle.sourceforge.net/config_javadoc.html#JavadocTagContinuationIndentation";>JavadocTagContinuationIndentation
     
     offset: "2"
    -779
    +780
      Error
     
     
     http://checkstyle.sourceforge.net/config_javadoc.html#NonEmptyAtclauseDescription";>NonEmptyAtclauseDescription
    -3209
    +3193
      Error
     
     misc
    @@ -8334,7 +8334,7 @@
     
     max: "100"
     ignorePattern: "^package.*|^import.*|a 
    href|href|http://|https://|ftp://|org.apache.thrift.|com.google.protobuf.|hbase.protobuf.generated"
    -1128
    +1125
      Error
     
     
    @@ -19397,103 +19397,91 @@
     javadoc
     NonEmptyAtclauseDescription
     At-clause should have a non-empty description.
    -570
    +569
     
      Error
     indentation
     Indentation
     'method def modifier' have incorrect indentation level 10, expected level 
    should be one of the following: 6, 8.
    -578
    +577
     
      Error
     indentation
     Indentation
     'method def' child have incorrect indentation level 12, expected level 
    should be one of the following: 8, 10.
    -580
    +579
     
      Error
     indentation
     Indentation
     'method def' child have incorrect indentation level 12, expected level 
    should be one of the following: 8, 10.
    -583
    +582
     
      Error

    [49/51] [partial] hbase-site git commit: Published site at .

    2017-10-27 Thread git-site-role
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/deprecated-list.html
    --
    diff --git a/apidocs/deprecated-list.html b/apidocs/deprecated-list.html
    index 078f646..8124883 100644
    --- a/apidocs/deprecated-list.html
    +++ b/apidocs/deprecated-list.html
    @@ -595,662 +595,768 @@
     org.apache.hadoop.hbase.rest.client.RemoteHTable.exists(List)
     
     
    +org.apache.hadoop.hbase.filter.FilterList.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.WhileMatchFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.PageFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.MultipleColumnPrefixFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.InclusiveStopFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.KeyOnlyFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.RowFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.ColumnRangeFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.FamilyFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.RandomRowFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.FirstKeyValueMatchingQualifiersFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.SkipFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.DependentColumnFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.Filter.filterKeyValue(Cell)
    +As of release 2.0.0, this 
    will be removed in HBase 3.0.0.
    + Instead use filterCell(Cell)
    +
    +
    +
    +org.apache.hadoop.hbase.filter.ColumnPaginationFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.ValueFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.ColumnCountGetFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.QualifierFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.PrefixFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.FuzzyRowFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.TimestampsFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.ColumnPrefixFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.MultiRowRangeFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.SingleColumnValueFilter.filterKeyValue(Cell)
    +
    +
    +org.apache.hadoop.hbase.filter.FirstKeyOnlyFilter.filterKeyValue(Cell)
    +
    +
     org.apache.hadoop.hbase.filter.Filter.filterRowKey(byte[],
     int, int)
     As of release 2.0.0, this 
    will be removed in HBase 3.0.0.
      Instead use Filter.filterRowKey(Cell)
     
     
    -
    +
     org.apache.hadoop.hbase.client.Admin.getAlterStatus(byte[])
     Since 2.0.0. Will be 
    removed in 3.0.0. No longer needed now you get a Future
      on an operation.
     
     
    -
    +
     org.apache.hadoop.hbase.client.Admin.getAlterStatus(TableName)
     Since 2.0.0. Will be 
    removed in 3.0.0. No longer needed now you get a Future
      on an operation.
     
     
    -
    +
     org.apache.hadoop.hbase.HTableDescriptor.getColumnFamilies()
     
    -
    +
     org.apache.hadoop.hbase.HColumnDescriptor.getCompactionCompression()
     As of release 2.0.0, this 
    will be removed in HBase 3.0.0
      (https://issues.apache.org/jira/browse/HBASE-13655";>HBASE-13655).
      Use HColumnDescriptor.getCompactionCompressionType().
     
     
    -
    +
     org.apache.hadoop.hbase.HRegionInfo.getComparator()
     Use 
    Region#getCellComparator().  deprecated for hbase 2.0, remove for hbase 
    3.0
     
     
    -
    +
     org.apache.hadoop.hbase.HColumnDescriptor.getCompression()
     As of release 2.0.0, this 
    will be removed in HBase 3.0.0
      (https://issues.apache.org/jira/browse/HBASE-13655";>HBASE-13655).
      Use HColumnDescriptor.getCompressionType().
     
     
    -
    +
     org.apache.hadoop.hbase.ClusterStatus.getDeadServers()
     As of release 2.0.0, this 
    will be removed in HBase 3.0.0
      (https://issues.apache.org/jira/browse/HBASE-13656";>HBASE-13656).
      Use ClusterStatus.getDeadServersSize().
     
     
    -
    +
     org.apache.hadoop.hbase.HRegionInfo.getDescriptiveNameFromRegionStateForDisplay(RegionState,
     Configuration)
     As of release 2.0.0, this 
    will be removed in HBase 3.0.0
      Use 
    RegionInfoDisplay#getDescriptiveNameFromRegionStateForDisplay(RegionState, 
    Configuration)
      over in hbase-server module.
     
     
    -
    +
     org.apache.hadoop.hbase.HRegionInfo.getEndKeyForDisplay(HRegionInfo,
     Configuration)
     As of release 2.0.0, this 
    will be removed in HBase 3.0.0
      Use RegionInfoDisplay#getEndKeyForDisplay(RegionInfo, 
    Configuration)
      over in hbase-server module.
     
     
    -
    +
     org.apache.hadoop.hbase.HTableDescriptor.getFamilies()
     Use HTableDescriptor.getColumnFamilies().
     
     
    -
    +
     org.apache.hadoop.hbase.HTableDescriptor.getFamiliesKeys()
     As of release 2.0.0, this 
    will be removed in HBase 3.0.0
      (https://issues.apache.org/jira/browse/HBASE-18008";>HBASE-18008).
      Use HTableDescriptor.getColumnFamilyNames().
     
     
    -
    +
     org.apache.hadoop.hbase.HTableDescriptor.getFamily(byte[])
     Use HTa

    [10/51] [partial] hbase-site git commit: Published site at .

    2017-10-27 Thread git-site-role
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/MasterCoprocessorEnvironment.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/MasterCoprocessorEnvironment.html
     
    b/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/MasterCoprocessorEnvironment.html
    index 36f4f97..e452864 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/MasterCoprocessorEnvironment.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/MasterCoprocessorEnvironment.html
    @@ -329,8 +329,7 @@
     
     
     default void
    -MasterObserver.postGetLocks(ObserverContext ctx,
    -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
     title="class or interface in java.util">List lockedResources)
    +MasterObserver.postGetLocks(ObserverContext ctx)
     Called after a getLocks request has been processed.
     
     
    @@ -343,8 +342,7 @@
     
     
     default void
    -MasterObserver.postGetProcedures(ObserverContext ctx,
    - http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
     title="class or interface in java.util">List> procList)
    +MasterObserver.postGetProcedures(ObserverContext ctx)
     Called after a getProcedures request has been 
    processed.
     
     
    @@ -407,9 +405,7 @@
     
     
     default void
    -MasterObserver.postLockHeartbeat(ObserverContext ctx,
    - LockProcedure proc,
    - boolean keepAlive)
    +MasterObserver.postLockHeartbeat(ObserverContext ctx)
     Called after heartbeat to a lock.
     
     
    @@ -508,11 +504,10 @@
     
     
     default void
    -MasterObserver.postRequestLock(ObserverContext ctx,
    +MasterObserver.postRequestLock(ObserverContext ctx,
    http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String namespace,
    TableName tableName,
    RegionInfo[] regionInfos,
    -   LockType type,
    http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String description)
     Called after new LockProcedure is queued.
     
    @@ -635,8 +630,7 @@
     
     
     default void
    -MasterObserver.preAbortProcedure(ObserverContext ctx,
    - ProcedureExecutor procEnv,
    +MasterObserver.preAbortProcedure(ObserverContext ctx,
      long procId)
     Called before a abortProcedure request has been 
    processed.
     
    @@ -884,9 +878,9 @@
     
     
     default void
    -MasterObserver.preLockHeartbeat(ObserverContext ctx,
    -LockProcedure proc,
    -boolean keepAlive)
    +MasterObserver.preLockHeartbeat(ObserverContext ctx,
    +TableName tn,
    +http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String description)
     Called before heartbeat to a lock.
     
     
    @@ -1006,11 +1000,10 @@
     
     
     default void
    -MasterObserver.preRequestLock(ObserverContext ctx,
    +MasterObserver.preRequestLock(ObserverContext ctx,
       http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String namespace,
       TableName tableName,
       RegionInfo[] regionInfos,
    -  LockType type,
       http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String description)
     Called before new LockProcedure is queued.
     
    @@ -1325,100 +1318,94 @@
     
     
     void
    -AccessController.postGetProcedures(ObserverContext ctx,
    - http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
     title="class or interface in java.util">List> procList) 
    -
    -
    -void
     AccessController.postGetTableDescriptors(ObserverContext ctx,
    http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
     title="class or interface in java.util">List tableNamesList,
    http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
     title="class or interface in java.util">List descriptors,
    http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in 
    java.lang">String regex) 
     
    -
    +
     void
     Acces

    [12/51] [partial] hbase-site git commit: Published site at .

    2017-10-27 Thread git-site-role
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/devapidocs/org/apache/hadoop/hbase/client/Table.html
    --
    diff --git a/devapidocs/org/apache/hadoop/hbase/client/Table.html 
    b/devapidocs/org/apache/hadoop/hbase/client/Table.html
    index 0f1320b..b77757a 100644
    --- a/devapidocs/org/apache/hadoop/hbase/client/Table.html
    +++ b/devapidocs/org/apache/hadoop/hbase/client/Table.html
    @@ -18,7 +18,7 @@
     catch(err) {
     }
     //-->
    -var methods = 
    {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":38,"i7":6,"i8":38,"i9":6,"i10":6,"i11":38,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6,"i20":6,"i21":6,"i22":6,"i23":6,"i24":6,"i25":6,"i26":6,"i27":6,"i28":38,"i29":6,"i30":6,"i31":6,"i32":38,"i33":6,"i34":6,"i35":6,"i36":6,"i37":6,"i38":6,"i39":6,"i40":38,"i41":38,"i42":38,"i43":38};
    +var methods = 
    {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":38,"i7":6,"i8":38,"i9":6,"i10":6,"i11":38,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6,"i20":6,"i21":6,"i22":6,"i23":6,"i24":6,"i25":6,"i26":38,"i27":6,"i28":38,"i29":6,"i30":38,"i31":6,"i32":6,"i33":6,"i34":6,"i35":38,"i36":38,"i37":6,"i38":6,"i39":6,"i40":6,"i41":6,"i42":6,"i43":6,"i44":38,"i45":38,"i46":38,"i47":38};
     var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
    Methods"],4:["t3","Abstract Methods"],32:["t6","Deprecated Methods"]};
     var altColor = "altColor";
     var rowColor = "rowColor";
    @@ -110,7 +110,7 @@ var activeTableTab = "activeTableTab";
     
     
     @InterfaceAudience.Public
    -public interface Table
    +public interface Table
     extends http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true";
     title="class or interface in java.io">Closeable
     Used to communicate with a single HBase table.
      Obtain an instance from a Connection 
    and call close()
     afterwards.
    @@ -383,44 +383,69 @@ extends http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html
     
     int
     getOperationTimeout()
    -Get timeout (millisecond) of each operation for in Table 
    instance.
    +Deprecated. 
    +since 2.0 and will be 
    removed in 3.0 version
    + use getOperationTimeout(TimeUnit)
     instead
    +
     
     
     
    +long
    +getOperationTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
    +Get timeout of each operation in Table instance.
    +
    +
    +
     int
     getReadRpcTimeout()
    -Get timeout (millisecond) of each rpc read request in this 
    Table instance.
    +Deprecated. 
    +since 2.0 and will be 
    removed in 3.0 version
    + use getReadRpcTimeout(TimeUnit)
     instead
    +
     
     
    -
    +
    +long
    +getReadRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
    +Get timeout of each rpc read request in this Table 
    instance.
    +
    +
    +
     int
     getRpcTimeout()
     Deprecated. 
    -Use getReadRpcTimeout or 
    getWriteRpcTimeout instead
    +use getReadRpcTimeout(TimeUnit)
     or
    + getWriteRpcTimeout(TimeUnit)
     instead
     
     
     
    -
    +
    +long
    +getRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
    +Get timeout of each rpc request in this Table 
    instance.
    +
    +
    +
     ResultScanner
     getScanner(byte[] family)
     Gets a scanner on the current table for the given 
    family.
     
     
    -
    +
     ResultScanner
     getScanner(byte[] family,
       byte[] qualifier)
     Gets a scanner on the current table for the given family 
    and qualifier.
     
     
    -
    +
     ResultScanner
     getScanner(Scan scan)
     Returns a scanner on the current table as specified by the 
    Scan
      object.
     
     
    -
    +
     HTableDescriptor
     getTableDescriptor()
     Deprecated. 
    @@ -429,19 +454,28 @@ extends http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html
     
     
     
    -
    +
     int
     getWriteRpcTimeout()
    -Get timeout (millisecond) of each rpc write request in this 
    Table instance.
    +Deprecated. 
    +since 2.0 and will be 
    removed in 3.0 version
    + use getWriteRpcTimeout(TimeUnit)
     instead
    +
     
     
    -
    +
    +long
    +getWriteRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
    +Get timeout of each rpc write request in this Table 
    instance.
    +
    +
    +
     Result
     increment(Increment increment)
     Increments one or more columns within a single row.
     
     
    -
    +
     long
     incrementColumnValue(byte[] row,
     byte[] family,
    @@ -450,7 +484,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html
     See incrementColumnValue(byte[],
     byte[], byte[], long, Durability)
     
     
    -
    +
     long
     incrementColumnValue(byte[] row,
     byte[] family,
    @@ -460,25 +494,25 @@ extends http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html
     Atomically increments a column value.
     
     
    -
    +
     void
     mutateRow(RowMutations rm)
     Performs mult

    [25/51] [partial] hbase-site git commit: Published site at .

    2017-10-27 Thread git-site-role
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/src-html/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html
    --
    diff --git 
    a/apidocs/src-html/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html
     
    b/apidocs/src-html/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html
    index e8c4959..5856e0d 100644
    --- 
    a/apidocs/src-html/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html
    +++ 
    b/apidocs/src-html/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.html
    @@ -74,139 +74,145 @@
     066return false;
     067  }
     068
    -069  @Override
    -070  public ReturnCode filterKeyValue(Cell 
    kv) {
    -071if (sortedPrefixes.isEmpty()) {
    -072  return ReturnCode.INCLUDE;
    -073} else {
    -074  return filterColumn(kv);
    -075}
    -076  }
    -077
    -078  public ReturnCode filterColumn(Cell 
    cell) {
    -079byte [] qualifier = 
    CellUtil.cloneQualifier(cell);
    -080TreeSet 
    lesserOrEqualPrefixes =
    -081  (TreeSet) 
    sortedPrefixes.headSet(qualifier, true);
    -082
    -083if (lesserOrEqualPrefixes.size() != 
    0) {
    -084  byte [] 
    largestPrefixSmallerThanQualifier = lesserOrEqualPrefixes.last();
    -085  
    -086  if (Bytes.startsWith(qualifier, 
    largestPrefixSmallerThanQualifier)) {
    -087return ReturnCode.INCLUDE;
    -088  }
    -089  
    -090  if (lesserOrEqualPrefixes.size() == 
    sortedPrefixes.size()) {
    -091return ReturnCode.NEXT_ROW;
    -092  } else {
    -093hint = 
    sortedPrefixes.higher(largestPrefixSmallerThanQualifier);
    -094return 
    ReturnCode.SEEK_NEXT_USING_HINT;
    -095  }
    -096} else {
    -097  hint = sortedPrefixes.first();
    -098  return 
    ReturnCode.SEEK_NEXT_USING_HINT;
    -099}
    -100  }
    -101
    -102  public static Filter 
    createFilterFromArguments(ArrayList filterArguments) {
    -103byte [][] prefixes = new byte 
    [filterArguments.size()][];
    -104for (int i = 0 ; i < 
    filterArguments.size(); i++) {
    -105  byte [] columnPrefix = 
    ParseFilter.removeQuotesFromByteArray(filterArguments.get(i));
    -106  prefixes[i] = columnPrefix;
    -107}
    -108return new 
    MultipleColumnPrefixFilter(prefixes);
    -109  }
    -110
    -111  /**
    -112   * @return The filter serialized using 
    pb
    -113   */
    -114  public byte [] toByteArray() {
    -115
    FilterProtos.MultipleColumnPrefixFilter.Builder builder =
    -116  
    FilterProtos.MultipleColumnPrefixFilter.newBuilder();
    -117for (byte [] element : 
    sortedPrefixes) {
    -118  if (element != null) 
    builder.addSortedPrefixes(UnsafeByteOperations.unsafeWrap(element));
    -119}
    -120return 
    builder.build().toByteArray();
    -121  }
    -122
    -123  /**
    -124   * @param pbBytes A pb serialized 
    {@link MultipleColumnPrefixFilter} instance
    -125   * @return An instance of {@link 
    MultipleColumnPrefixFilter} made from bytes
    -126   * @throws DeserializationException
    -127   * @see #toByteArray
    -128   */
    -129  public static 
    MultipleColumnPrefixFilter parseFrom(final byte [] pbBytes)
    -130  throws DeserializationException {
    -131
    FilterProtos.MultipleColumnPrefixFilter proto;
    -132try {
    -133  proto = 
    FilterProtos.MultipleColumnPrefixFilter.parseFrom(pbBytes);
    -134} catch 
    (InvalidProtocolBufferException e) {
    -135  throw new 
    DeserializationException(e);
    -136}
    -137int numPrefixes = 
    proto.getSortedPrefixesCount();
    -138byte [][] prefixes = new 
    byte[numPrefixes][];
    -139for (int i = 0; i < numPrefixes; 
    ++i) {
    -140  prefixes[i] = 
    proto.getSortedPrefixes(i).toByteArray();
    -141}
    -142
    -143return new 
    MultipleColumnPrefixFilter(prefixes);
    -144  }
    -145
    -146  /**
    -147   * @param other
    -148   * @return true if and only if the 
    fields of the filter that are serialized
    -149   * are equal to the corresponding 
    fields in other.  Used for testing.
    -150   */
    -151  boolean areSerializedFieldsEqual(Filter 
    o) {
    -152if (o == this) return true;
    -153if (!(o instanceof 
    MultipleColumnPrefixFilter)) return false;
    -154
    -155MultipleColumnPrefixFilter other = 
    (MultipleColumnPrefixFilter)o;
    -156return 
    this.sortedPrefixes.equals(other.sortedPrefixes);
    -157  }
    -158
    -159  @Override
    -160  public Cell getNextCellHint(Cell cell) 
    {
    -161return 
    CellUtil.createFirstOnRowCol(cell, hint, 0, hint.length);
    -162  }
    -163
    -164  public TreeSet 
    createTreeSet() {
    -165return new TreeSet<>(new 
    Comparator() {
    -166@Override
    -167  public int compare (Object o1, 
    Object o2) {
    -168  if (o1 == null || o2 == null)
    -169throw new 
    IllegalArgumentException ("prefixes can't be null");
    -170
    -171  byte [] b1 = (byte []) o1;
    -172  byte [] b2 = (byte []) o2;
    -173  return Bytes.compareTo (b1, 0, 
    b1.length, b2, 0, b2.length);
    -174}
    -175  });
    -176  }
    -177
    -178  @Override
    -179  public String toString() {
    -180return toString(MAX_LOG_PREFIXES);
    -181  }
    -182
    -183  protected String toString(int 
    
    

    [41/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/org/apache/hadoop/hbase/filter/InclusiveStopFilter.html
    --
    diff --git a/apidocs/org/apache/hadoop/hbase/filter/InclusiveStopFilter.html 
    b/apidocs/org/apache/hadoop/hbase/filter/InclusiveStopFilter.html
    index 64cbea3..b1aefe7 100644
    --- a/apidocs/org/apache/hadoop/hbase/filter/InclusiveStopFilter.html
    +++ b/apidocs/org/apache/hadoop/hbase/filter/InclusiveStopFilter.html
    @@ -18,8 +18,8 @@
     catch(err) {
     }
     //-->
    -var methods = {"i0":9,"i1":10,"i2":10,"i3":10,"i4":10,"i5":9,"i6":10,"i7":10};
    -var tabs = {65535:["t0","All Methods"],1:["t1","Static 
    Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
    +var methods = 
    {"i0":9,"i1":10,"i2":10,"i3":42,"i4":10,"i5":10,"i6":9,"i7":10,"i8":10};
    +var tabs = {65535:["t0","All Methods"],1:["t1","Static 
    Methods"],2:["t2","Instance Methods"],8:["t4","Concrete 
    Methods"],32:["t6","Deprecated Methods"]};
     var altColor = "altColor";
     var rowColor = "rowColor";
     var tableTab = "tableTab";
    @@ -186,7 +186,7 @@ extends org.apache.hadoop.hbase.filter.FilterBase
     
     Method Summary
     
    -All Methods Static Methods Instance Methods Concrete Methods 
    +All Methods Static Methods Instance Methods Concrete Methods Deprecated Methods 
     
     Modifier and Type
     Method and Description
    @@ -204,31 +204,37 @@ extends org.apache.hadoop.hbase.filter.FilterBase
     
     
     Filter.ReturnCode
    -filterKeyValue(Cell v)
    +filterCell(Cell c)
     A way to filter based on the column family, column 
    qualifier and/or the column value.
     
     
     
    +Filter.ReturnCode
    +filterKeyValue(Cell c)
    +Deprecated. 
    +
    +
    +
     boolean
     filterRowKey(Cell firstRowCell)
     Filters a row based on the row key.
     
     
    -
    +
     byte[]
     getStopRowKey() 
     
    -
    +
     static InclusiveStopFilter
     parseFrom(byte[] pbBytes) 
     
    -
    +
     byte[]
     toByteArray()
     Return length 0 byte array for Filters that don't require 
    special serialization
     
     
    -
    +
     http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String
     toString()
     Return filter's info for debugging and logging 
    purpose.
    @@ -302,7 +308,9 @@ extends org.apache.hadoop.hbase.filter.FilterBase
     
     
     filterKeyValue
    -public Filter.ReturnCode filterKeyValue(Cell v)
    +http://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true";
     title="class or interface in java.lang">@Deprecated
    +public Filter.ReturnCode filterKeyValue(Cell c)
    +Deprecated. 
     Description copied from 
    class: Filter
     A way to filter based on the column family, column 
    qualifier and/or the column value. Return
      code is described below. This allows filters to filter only certain number of 
    columns, then
    @@ -315,14 +323,48 @@ extends org.apache.hadoop.hbase.filter.FilterBase
      If your filter returns ReturnCode.NEXT_ROW, it should return
      ReturnCode.NEXT_ROW until Filter.reset()
     is called just in case the caller calls
      for the next row.
    - 
    +
      Concrete implementers can signal a failure condition in their code by 
    throwing an
      http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
     title="class or interface in java.io">IOException.
     
    -Specified by:
    +Overrides:
     filterKeyValue in
     class Filter
     Parameters:
    -v - the Cell in question
    +c - the Cell in question
    +Returns:
    +code as described below, Filter.ReturnCode.INCLUDE by default
    +See Also:
    +Filter.ReturnCode
    +
    +
    +
    +
    +
    +
    +
    +
    +filterCell
    +public Filter.ReturnCode filterCell(Cell c)
    +Description copied from 
    class: Filter
    +A way to filter based on the column family, column 
    qualifier and/or the column value. Return
    + code is described below. This allows filters to filter only certain number of 
    columns, then
    + terminate without matching ever column.
    +
    + If filterRowKey returns true, filterCell needs to be consistent with it.
    +
    + filterCell can assume that filterRowKey has already been called for the row.
    +
    + If your filter returns ReturnCode.NEXT_ROW, it should return
    + ReturnCode.NEXT_ROW until Filter.reset()
     is called just in case the caller calls
    + for the next row.
    +
    + Concrete implementers can signal a failure condition in their code by 
    throwing an
    + http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
     title="class or interface in java.io">IOException.
    +
    +Overrides:
    +filterCell in
     class Filter
    +Parameters:
    +c - the Cell in question
     Returns:
     code as described below
     See Also:
    @@ -336,10 +378,10 @@ extends org.apache.hadoop.hbase.filter.FilterBase
     
     
     filterRowKey
    -public boolean filterRowKey(Cell firstRowCell)
    +public boolean filterRowKey(Cell firstRowCell)
     Description copied from 
    class: Filter
     Filters a row based on the row key. If this returns true, 
    the entire row will be excluded. If
    - false, each KeyValue in the row will be passed to Filter.filterKeyValue(Cell)
     below.
    + false, each KeyValue in the row will be passed to Filter.filterCell(Cell)
     below.
      If Filter.filterAllRemaining()
     return

    hbase-site git commit: INFRA-10751 Empty commit

    Repository: hbase-site
    Updated Branches:
      refs/heads/asf-site 00c223882 -> 8021fef8a
    
    
    INFRA-10751 Empty commit
    
    
    Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo
    Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/8021fef8
    Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/8021fef8
    Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/8021fef8
    
    Branch: refs/heads/asf-site
    Commit: 8021fef8a4bbbd22a2d320ff8eb66e7e25c91dbd
    Parents: 00c2238
    Author: jenkins 
    Authored: Fri Oct 27 15:16:42 2017 +
    Committer: jenkins 
    Committed: Fri Oct 27 15:16:42 2017 +
    
    --
    
    --
    
    
    
    

    [45/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/org/apache/hadoop/hbase/client/Table.html
    --
    diff --git a/apidocs/org/apache/hadoop/hbase/client/Table.html 
    b/apidocs/org/apache/hadoop/hbase/client/Table.html
    index bc44d99..08c2075 100644
    --- a/apidocs/org/apache/hadoop/hbase/client/Table.html
    +++ b/apidocs/org/apache/hadoop/hbase/client/Table.html
    @@ -18,7 +18,7 @@
     catch(err) {
     }
     //-->
    -var methods = 
    {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":38,"i7":6,"i8":38,"i9":6,"i10":6,"i11":38,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6,"i20":6,"i21":6,"i22":6,"i23":6,"i24":6,"i25":6,"i26":6,"i27":6,"i28":38,"i29":6,"i30":6,"i31":6,"i32":38,"i33":6,"i34":6,"i35":6,"i36":6,"i37":6,"i38":6,"i39":6,"i40":38,"i41":38,"i42":38,"i43":38};
    +var methods = 
    {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":6,"i6":38,"i7":6,"i8":38,"i9":6,"i10":6,"i11":38,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6,"i20":6,"i21":6,"i22":6,"i23":6,"i24":6,"i25":6,"i26":38,"i27":6,"i28":38,"i29":6,"i30":38,"i31":6,"i32":6,"i33":6,"i34":6,"i35":38,"i36":38,"i37":6,"i38":6,"i39":6,"i40":6,"i41":6,"i42":6,"i43":6,"i44":38,"i45":38,"i46":38,"i47":38};
     var tabs = {65535:["t0","All Methods"],2:["t2","Instance 
    Methods"],4:["t3","Abstract Methods"],32:["t6","Deprecated Methods"]};
     var altColor = "altColor";
     var rowColor = "rowColor";
    @@ -110,7 +110,7 @@ var activeTableTab = "activeTableTab";
     
     
     @InterfaceAudience.Public
    -public interface Table
    +public interface Table
     extends http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html?is-external=true";
     title="class or interface in java.io">Closeable
     Used to communicate with a single HBase table.
      Obtain an instance from a Connection 
    and call close()
     afterwards.
    @@ -383,44 +383,69 @@ extends http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html
     
     int
     getOperationTimeout()
    -Get timeout (millisecond) of each operation for in Table 
    instance.
    +Deprecated. 
    +since 2.0 and will be 
    removed in 3.0 version
    + use getOperationTimeout(TimeUnit)
     instead
    +
     
     
     
    +long
    +getOperationTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
    +Get timeout of each operation in Table instance.
    +
    +
    +
     int
     getReadRpcTimeout()
    -Get timeout (millisecond) of each rpc read request in this 
    Table instance.
    +Deprecated. 
    +since 2.0 and will be 
    removed in 3.0 version
    + use getReadRpcTimeout(TimeUnit)
     instead
    +
     
     
    -
    +
    +long
    +getReadRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
    +Get timeout of each rpc read request in this Table 
    instance.
    +
    +
    +
     int
     getRpcTimeout()
     Deprecated. 
    -Use getReadRpcTimeout or 
    getWriteRpcTimeout instead
    +use getReadRpcTimeout(TimeUnit)
     or
    + getWriteRpcTimeout(TimeUnit)
     instead
     
     
     
    -
    +
    +long
    +getRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
    +Get timeout of each rpc request in this Table 
    instance.
    +
    +
    +
     ResultScanner
     getScanner(byte[] family)
     Gets a scanner on the current table for the given 
    family.
     
     
    -
    +
     ResultScanner
     getScanner(byte[] family,
       byte[] qualifier)
     Gets a scanner on the current table for the given family 
    and qualifier.
     
     
    -
    +
     ResultScanner
     getScanner(Scan scan)
     Returns a scanner on the current table as specified by the 
    Scan
      object.
     
     
    -
    +
     HTableDescriptor
     getTableDescriptor()
     Deprecated. 
    @@ -429,19 +454,28 @@ extends http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html
     
     
     
    -
    +
     int
     getWriteRpcTimeout()
    -Get timeout (millisecond) of each rpc write request in this 
    Table instance.
    +Deprecated. 
    +since 2.0 and will be 
    removed in 3.0 version
    + use getWriteRpcTimeout(TimeUnit)
     instead
    +
     
     
    -
    +
    +long
    +getWriteRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
    +Get timeout of each rpc write request in this Table 
    instance.
    +
    +
    +
     Result
     increment(Increment increment)
     Increments one or more columns within a single row.
     
     
    -
    +
     long
     incrementColumnValue(byte[] row,
     byte[] family,
    @@ -450,7 +484,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html
     See incrementColumnValue(byte[],
     byte[], byte[], long, Durability)
     
     
    -
    +
     long
     incrementColumnValue(byte[] row,
     byte[] family,
    @@ -460,25 +494,25 @@ extends http://docs.oracle.com/javase/8/docs/api/java/io/Closeable.html
     Atomically increments a column value.
     
     
    -
    +
     void
     mutateRow(RowMutations rm)
     Performs multiple mutations 

    [13/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/devapidocs/org/apache/hadoop/hbase/client/HTable.html
    --
    diff --git a/devapidocs/org/apache/hadoop/hbase/client/HTable.html 
    b/devapidocs/org/apache/hadoop/hbase/client/HTable.html
    index 26e7b5a..b21044b 100644
    --- a/devapidocs/org/apache/hadoop/hbase/client/HTable.html
    +++ b/devapidocs/org/apache/hadoop/hbase/client/HTable.html
    @@ -18,7 +18,7 @@
     catch(err) {
     }
     //-->
    -var methods = 
    {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":9,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":9,"i33":10,"i34":10,"i35":10,"i36":9,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":42,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":42,"i57":42,"i58":42,"i59":42,"i60":10,"i61":10,"i62":9};
    +var methods = 
    {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":9,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":9,"i33":10,"i34":10,"i35":10,"i36":9,"i37":10,"i38":42,"i39":10,"i40":10,"i41":42,"i42":10,"i43":10,"i44":42,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":42,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":42,"i61":42,"i62":42,"i63":42,"i64":10,"i65":10,"i66":9};
     var tabs = {65535:["t0","All Methods"],1:["t1","Static 
    Methods"],2:["t2","Instance Methods"],8:["t4","Concrete 
    Methods"],32:["t6","Deprecated Methods"]};
     var altColor = "altColor";
     var rowColor = "rowColor";
    @@ -199,7 +199,7 @@ implements 
     
     private int
    -operationTimeout 
    +operationTimeoutMs 
     
     
     private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">ExecutorService
    @@ -207,7 +207,7 @@ implements 
     
     private int
    -readRpcTimeout 
    +readRpcTimeoutMs 
     
     
     private RpcRetryingCallerFactory
    @@ -219,7 +219,7 @@ implements 
     
     private int
    -rpcTimeout 
    +rpcTimeoutMs 
     
     
     private int
    @@ -235,7 +235,7 @@ implements 
     
     private int
    -writeRpcTimeout 
    +writeRpcTimeoutMs 
     
     
     
    @@ -597,74 +597,98 @@ implements 
     int
     getOperationTimeout()
    -Get timeout (millisecond) of each operation for in Table 
    instance.
    +Deprecated. 
     
     
     
    +long
    +getOperationTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
    +Get timeout of each operation in Table instance.
    +
    +
    +
     (package private) http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">ExecutorService
     getPool()
     The pool is used for mutli requests for this HTable
     
     
    -
    +
     int
     getReadRpcTimeout()
    -Get timeout (millisecond) of each rpc read request in this 
    Table instance.
    +Deprecated. 
     
     
    -
    +
    +long
    +getReadRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
    +Get timeout of each rpc read request in this Table 
    instance.
    +
    +
    +
     RegionLocator
     getRegionLocator() 
     
    -
    +
     int
     getRpcTimeout()
     Deprecated. 
     
     
    -
    +
    +long
    +getRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
    +Get timeout of each rpc request in this Table 
    instance.
    +
    +
    +
     ResultScanner
     getScanner(byte[] family)
     The underlying HTable must not be 
    closed.
     
     
    -
    +
     ResultScanner
     getScanner(byte[] family,
       byte[] qualifier)
     The underlying HTable must not be 
    closed.
     
     
    -
    +
     ResultScanner
     getScanner(Scan scan)
     The underlying HTable must not be 
    closed.
     
     
    -
    +
     private http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
     title="class or interface in java.util">List
     getStartKeysInRange(byte[] start,
    byte[] end) 
     
    -
    +
     HTableDescriptor
     getTableDescriptor()
     Gets the table descriptor for 
    this table.
     
     
    -
    +
     int
     getWriteRpcTimeout()
    -Get timeout (millisecond) of each rpc write request in this 
    Table instance.
    +Deprecated. 
     
     
    -
    +
    +long
    +getWriteRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
    +Get timeout of each rpc write request in this Table 
    instance.
    +
    +
    +
     Result
     increment(Increment increment)
     Increments one or more columns within a single row.
     
     
    -
    +
     long
     incrementColumnValue(byte[] row,
    

    [09/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html 
    b/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
    index 15dd9d4..e7584be 100644
    --- 
    a/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
    +++ 
    b/devapidocs/org/apache/hadoop/hbase/coprocessor/class-use/ObserverContext.html
    @@ -612,8 +612,7 @@
     
     
     default void
    -MasterObserver.postGetLocks(ObserverContext ctx,
    -http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
     title="class or interface in java.util">List lockedResources)
    +MasterObserver.postGetLocks(ObserverContext ctx)
     Called after a getLocks request has been processed.
     
     
    @@ -634,8 +633,7 @@
     
     
     default void
    -MasterObserver.postGetProcedures(ObserverContext ctx,
    - http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
     title="class or interface in java.util">List> procList)
    +MasterObserver.postGetProcedures(ObserverContext ctx)
     Called after a getProcedures request has been 
    processed.
     
     
    @@ -715,9 +713,7 @@
     
     
     default void
    -MasterObserver.postLockHeartbeat(ObserverContext ctx,
    - LockProcedure proc,
    - boolean keepAlive)
    +MasterObserver.postLockHeartbeat(ObserverContext ctx)
     Called after heartbeat to a lock.
     
     
    @@ -865,11 +861,10 @@
     
     
     default void
    -MasterObserver.postRequestLock(ObserverContext ctx,
    +MasterObserver.postRequestLock(ObserverContext ctx,
    http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String namespace,
    TableName tableName,
    RegionInfo[] regionInfos,
    -   LockType type,
    http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String description)
     Called after new LockProcedure is queued.
     
    @@ -1088,8 +1083,7 @@
     
     
     default void
    -MasterObserver.preAbortProcedure(ObserverContext ctx,
    - ProcedureExecutor procEnv,
    +MasterObserver.preAbortProcedure(ObserverContext ctx,
      long procId)
     Called before a abortProcedure request has been 
    processed.
     
    @@ -1530,9 +1524,9 @@
     
     
     default void
    -MasterObserver.preLockHeartbeat(ObserverContext ctx,
    -LockProcedure proc,
    -boolean keepAlive)
    +MasterObserver.preLockHeartbeat(ObserverContext ctx,
    +TableName tn,
    +http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String description)
     Called before heartbeat to a lock.
     
     
    @@ -1700,11 +1694,10 @@
     
     
     default void
    -MasterObserver.preRequestLock(ObserverContext ctx,
    +MasterObserver.preRequestLock(ObserverContext ctx,
       http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String namespace,
       TableName tableName,
       RegionInfo[] regionInfos,
    -  LockType type,
       http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String description)
     Called before new LockProcedure is queued.
     
    @@ -2218,38 +2211,33 @@
     
     
     void
    -AccessController.postGetProcedures(ObserverContext ctx,
    - http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
     title="class or interface in java.util">List> procList) 
    -
    -
    -void
     AccessController.postGetTableDescriptors(ObserverContext ctx,
    http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
     title="class or interface in java.util">List tableNamesList,
    http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
     title="class or interface in java.util">List descriptors,
    http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in 
    java.lang">String regex) 
     
    -
    +
     void
     AccessController.postGetTableNames(ObserverContext
    

    [16/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
    --
    diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html 
    b/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
    index 10a1e43..d553f00 100644
    --- a/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
    +++ b/devapidocs/org/apache/hadoop/hbase/class-use/Cell.html
    @@ -2883,7 +2883,7 @@ service.
     
     protected Cell
     FilterListBase.transformedCell
    -When filtering a given Cell in FilterListBase.filterKeyValue(Cell),
     this stores the transformed Cell
    +When filtering a given Cell in FilterListBase.filterCell(Cell),
     this stores the transformed Cell
      to be returned by FilterListBase.transformCell(Cell).
     
     
    @@ -2919,60 +2919,60 @@ service.
     
     
     Cell
    -FilterWrapper.getNextCellHint(Cell currentCell) 
    -
    -
    -Cell
     FilterBase.getNextCellHint(Cell currentCell)
     Filters that are not sure which key must be next seeked to, 
    can inherit
      this implementation that, by default, returns a null Cell.
     
     
    -
    +
     Cell
     ColumnPrefixFilter.getNextCellHint(Cell cell) 
     
    -
    +
     Cell
     ColumnRangeFilter.getNextCellHint(Cell cell) 
     
    -
    +
     Cell
     MultipleColumnPrefixFilter.getNextCellHint(Cell cell) 
     
    -
    +
     Cell
     ColumnPaginationFilter.getNextCellHint(Cell cell) 
     
    -
    +
     Cell
     FilterListWithOR.getNextCellHint(Cell currentCell) 
     
    -
    +
     Cell
     MultiRowRangeFilter.getNextCellHint(Cell currentKV) 
     
    -
    +
     abstract Cell
     Filter.getNextCellHint(Cell currentCell)
     If the filter returns the match code SEEK_NEXT_USING_HINT, 
    then it should also tell which is
      the next key it must seek to.
     
     
    -
    +
     Cell
     TimestampsFilter.getNextCellHint(Cell currentCell)
     Pick the next cell that the scanner should seek to.
     
     
    -
    +
     Cell
     FuzzyRowFilter.getNextCellHint(Cell currentCell) 
     
    -
    +
     Cell
     FilterList.getNextCellHint(Cell currentCell) 
     
    +
    +Cell
    +FilterWrapper.getNextCellHint(Cell currentCell) 
    +
     
     Cell
     SkipFilter.transformCell(Cell v) 
    @@ -2983,34 +2983,34 @@ service.
     
     
     Cell
    -FilterWrapper.transformCell(Cell v) 
    -
    -
    -Cell
     FilterBase.transformCell(Cell v)
     By default no transformation takes place
     
      Give the filter a chance to transform the passed KeyValue.
     
     
    -
    +
     Cell
     KeyOnlyFilter.transformCell(Cell cell) 
     
    -
    +
     abstract Cell
     Filter.transformCell(Cell v)
     Give the filter a chance to transform the passed 
    KeyValue.
     
     
    -
    +
     Cell
     WhileMatchFilter.transformCell(Cell v) 
     
    -
    +
     Cell
     FilterList.transformCell(Cell c) 
     
    +
    +Cell
    +FilterWrapper.transformCell(Cell v) 
    +
     
     
     
    @@ -3105,139 +3105,298 @@ service.
     
     
     Filter.ReturnCode
    -ColumnPrefixFilter.filterColumn(Cell cell) 
    +ValueFilter.filterCell(Cell c) 
     
     
     Filter.ReturnCode
    -MultipleColumnPrefixFilter.filterColumn(Cell cell) 
    +SkipFilter.filterCell(Cell c) 
     
     
    -private boolean
    -SingleColumnValueFilter.filterColumnValue(Cell cell) 
    +Filter.ReturnCode
    +FilterListBase.filterCell(Cell c) 
     
     
     Filter.ReturnCode
    -ValueFilter.filterKeyValue(Cell v) 
    +FamilyFilter.filterCell(Cell c) 
     
     
     Filter.ReturnCode
    -SkipFilter.filterKeyValue(Cell v) 
    +ColumnPrefixFilter.filterCell(Cell cell) 
     
     
     Filter.ReturnCode
    -FilterListBase.filterKeyValue(Cell c) 
    +PageFilter.filterCell(Cell ignored) 
     
     
     Filter.ReturnCode
    -FamilyFilter.filterKeyValue(Cell v) 
    +RowFilter.filterCell(Cell v) 
     
     
     Filter.ReturnCode
    -FilterWrapper.filterKeyValue(Cell v) 
    +ColumnRangeFilter.filterCell(Cell c) 
     
     
     Filter.ReturnCode
    -ColumnPrefixFilter.filterKeyValue(Cell cell) 
    +ColumnCountGetFilter.filterCell(Cell c) 
     
     
     Filter.ReturnCode
    -PageFilter.filterKeyValue(Cell ignored) 
    +MultipleColumnPrefixFilter.filterCell(Cell c) 
     
     
     Filter.ReturnCode
    -RowFilter.filterKeyValue(Cell v) 
    +ColumnPaginationFilter.filterCell(Cell c) 
     
     
     Filter.ReturnCode
    -ColumnRangeFilter.filterKeyValue(Cell kv) 
    +DependentColumnFilter.filterCell(Cell c) 
     
     
     Filter.ReturnCode
    -ColumnCountGetFilter.filterKeyValue(Cell v) 
    +InclusiveStopFilter.filterCell(Cell c) 
     
     
     Filter.ReturnCode
    -MultipleColumnPrefixFilter.filterKeyValue(Cell kv) 
    +KeyOnlyFilter.filterCell(Cell ignored) 
     
     
     Filter.ReturnCode
    -ColumnPaginationFilter.filterKeyValue(Cell v) 
    +MultiRowRangeFilter.filterCell(Cell ignored) 
     
     
     Filter.ReturnCode
    -DependentColumnFilter.filterKeyValue(Cell c) 
    +Filter.filterCell(Cell c)
    +A way to filter based on the column family, column 
    qualifier and/or the column value.
    +
     
     
     Filter.ReturnCode
    -InclusiveStopFilter.filterKeyValue(Cell v) 
    +FirstKeyOnlyFilter.filterCell(Cell c) 
     
     
     Filter.ReturnCode
    -KeyOnlyFilter.filterKeyValue(Cell ignored) 
    +WhileMatchFilter.filterCell(Cell c) 
     
     
     Filter.ReturnCode
    -MultiRowRangeFilter.filterKeyValue(Cell ignored) 
    +FirstKeyValueMatchingQualifiersFilter.filterCell(Cell c)
    +Deprecated. 
    + 
     
     
    -abstract Filter.ReturnCode
    -Filter.filterKeyValue(Cell v)
    -A way to filter based on the column family, column 
    qualifier and/or the column value.
    +Filter.ReturnCode
    +TimestampsFilter.filterCell(Cell c) 
    +
    +
    +Filter.ReturnCode

    [01/51] [partial] hbase-site git commit: Published site at .

    Repository: hbase-site
    Updated Branches:
      refs/heads/asf-site 7bee80a09 -> 00c223882
    
    
    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/devapidocs/org/apache/hadoop/hbase/filter/KeyOnlyFilter.KeyOnlyCell.html
    --
    diff --git 
    a/devapidocs/org/apache/hadoop/hbase/filter/KeyOnlyFilter.KeyOnlyCell.html 
    b/devapidocs/org/apache/hadoop/hbase/filter/KeyOnlyFilter.KeyOnlyCell.html
    index a1c2674..c3d4b9c 100644
    --- a/devapidocs/org/apache/hadoop/hbase/filter/KeyOnlyFilter.KeyOnlyCell.html
    +++ b/devapidocs/org/apache/hadoop/hbase/filter/KeyOnlyFilter.KeyOnlyCell.html
    @@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab";
     
     
     
    -static class KeyOnlyFilter.KeyOnlyCell
    +static class KeyOnlyFilter.KeyOnlyCell
     extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true";
     title="class or interface in java.lang">Object
     implements Cell
     
    @@ -294,7 +294,7 @@ implements 
     
     cell
    -private Cell cell
    +private Cell cell
     
     
     
    @@ -303,7 +303,7 @@ implements 
     
     lenAsVal
    -private boolean lenAsVal
    +private boolean lenAsVal
     
     
     
    @@ -320,7 +320,7 @@ implements 
     
     KeyOnlyCell
    -public KeyOnlyCell(Cell c,
    +public KeyOnlyCell(Cell c,
    boolean lenAsVal)
     
     
    @@ -338,7 +338,7 @@ implements 
     
     getRowArray
    -public byte[] getRowArray()
    +public byte[] getRowArray()
     Description copied from 
    interface: Cell
     Contiguous raw bytes that may start at any index in the 
    containing array. Max length is
      Short.MAX_VALUE which is 32,767 bytes.
    @@ -356,7 +356,7 @@ implements 
     
     getRowOffset
    -public int getRowOffset()
    +public int getRowOffset()
     
     Specified by:
     getRowOffset in
     interface Cell
    @@ -371,7 +371,7 @@ implements 
     
     getRowLength
    -public short getRowLength()
    +public short getRowLength()
     
     Specified by:
     getRowLength in
     interface Cell
    @@ -386,7 +386,7 @@ implements 
     
     getFamilyArray
    -public byte[] getFamilyArray()
    +public byte[] getFamilyArray()
     Description copied from 
    interface: Cell
     Contiguous bytes composed of legal HDFS filename characters 
    which may start at any index in the
      containing array. Max length is Byte.MAX_VALUE, which is 127 bytes.
    @@ -404,7 +404,7 @@ implements 
     
     getFamilyOffset
    -public int getFamilyOffset()
    +public int getFamilyOffset()
     
     Specified by:
     getFamilyOffset in
     interface Cell
    @@ -419,7 +419,7 @@ implements 
     
     getFamilyLength
    -public byte getFamilyLength()
    +public byte getFamilyLength()
     
     Specified by:
     getFamilyLength in
     interface Cell
    @@ -434,7 +434,7 @@ implements 
     
     getQualifierArray
    -public byte[] getQualifierArray()
    +public byte[] getQualifierArray()
     Description copied from 
    interface: Cell
     Contiguous raw bytes that may start at any index in the 
    containing array.
     
    @@ -451,7 +451,7 @@ implements 
     
     getQualifierOffset
    -public int getQualifierOffset()
    +public int getQualifierOffset()
     
     Specified by:
     getQualifierOffset in
     interface Cell
    @@ -466,7 +466,7 @@ implements 
     
     getQualifierLength
    -public int getQualifierLength()
    +public int getQualifierLength()
     
     Specified by:
     getQualifierLength in
     interface Cell
    @@ -481,7 +481,7 @@ implements 
     
     getTimestamp
    -public long getTimestamp()
    +public long getTimestamp()
     
     Specified by:
     getTimestamp in
     interface Cell
    @@ -497,7 +497,7 @@ implements 
     
     getTypeByte
    -public byte getTypeByte()
    +public byte getTypeByte()
     
     Specified by:
     getTypeByte in
     interface Cell
    @@ -512,7 +512,7 @@ implements 
     
     getSequenceId
    -public long getSequenceId()
    +public long getSequenceId()
     Description copied from 
    interface: Cell
     A region-specific unique monotonically increasing sequence 
    ID given to each Cell. It always
      exists for cells in the memstore but is not retained forever. It will be kept 
    for
    @@ -532,7 +532,7 @@ implements 
     
     getValueArray
    -public byte[] getValueArray()
    +public byte[] getValueArray()
     Description copied from 
    interface: Cell
     Contiguous raw bytes that may start at any index in the 
    containing array. Max length is
      Integer.MAX_VALUE which is 2,147,483,647 bytes.
    @@ -550,7 +550,7 @@ implements 
     
     getValueOffset
    -public int getValueOffset()
    +public int getValueOffset()
     
     Specified by:
     getValueOffset in
     interface Cell
    @@ -565,7 +565,7 @@ implements 
     
     getValueLength
    -public int getValueLength()
    +public int getValueLength()
     
     Specified by:
     getValueLength in
     interface Cell
    @@ -580,7 +580,7 @@ implements 
     
     getTagsArray
    -public byte[] getTagsArray()
    +public byte[] getTagsArray()
     Description copied from 
    interface: Cell
     Contiguous raw bytes representing tags that may start at 
    any index in the containing array.
     
    @@ -597,7 +597,7 @@ implements 
     
     getTagsOffset
    -public int getTagsOffset()
    +public int getTagsOffset()
     
     Specified by:
     getTagsOffset in
     interface Cell
    @@ -612,7 +612,7 @@ implements 
     
     getTagsLength
    -public int getTagsLength()
    +public int getTagsLength()
     Description copied from 
    interface: Cell
     HBase internally uses 2 bytes to store tags length in Cell.
      As the tags length is alwa

    [43/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/org/apache/hadoop/hbase/filter/DependentColumnFilter.html
    --
    diff --git a/apidocs/org/apache/hadoop/hbase/filter/DependentColumnFilter.html 
    b/apidocs/org/apache/hadoop/hbase/filter/DependentColumnFilter.html
    index 35033a5..5e49fb0 100644
    --- a/apidocs/org/apache/hadoop/hbase/filter/DependentColumnFilter.html
    +++ b/apidocs/org/apache/hadoop/hbase/filter/DependentColumnFilter.html
    @@ -18,8 +18,8 @@
     catch(err) {
     }
     //-->
    -var methods = 
    {"i0":9,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":9,"i12":10,"i13":10,"i14":10};
    -var tabs = {65535:["t0","All Methods"],1:["t1","Static 
    Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
    +var methods = 
    {"i0":9,"i1":10,"i2":10,"i3":10,"i4":42,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":9,"i13":10,"i14":10,"i15":10};
    +var tabs = {65535:["t0","All Methods"],1:["t1","Static 
    Methods"],2:["t2","Instance Methods"],8:["t4","Concrete 
    Methods"],32:["t6","Deprecated Methods"]};
     var altColor = "altColor";
     var rowColor = "rowColor";
     var tableTab = "tableTab";
    @@ -264,7 +264,7 @@ extends 
    -All Methods Static Methods Instance Methods Concrete Methods 
    +All Methods Static Methods Instance Methods Concrete Methods Deprecated Methods 
     
     Modifier and Type
     Method and Description
    @@ -286,26 +286,32 @@ extends 
     Filter.ReturnCode
    -filterKeyValue(Cell c)
    +filterCell(Cell c)
     A way to filter based on the column family, column 
    qualifier and/or the column value.
     
     
     
    +Filter.ReturnCode
    +filterKeyValue(Cell c)
    +Deprecated. 
    +
    +
    +
     boolean
     filterRow()
     Filters that never filter by rows based on previously 
    gathered state from
    - Filter.filterKeyValue(Cell)
     can inherit this implementation that
    + Filter.filterCell(Cell)
     can inherit this implementation that
      never filters a row.
     
     
    -
    +
     void
     filterRowCells(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
     title="class or interface in java.util">List kvs)
     Filters that never filter by modifying the returned List of 
    Cells can
      inherit this implementation that does nothing.
     
     
    -
    +
     boolean
     filterRowKey(byte[] buffer,
     int offset,
    @@ -314,43 +320,43 @@ extends 
    +
     boolean
     getDropDependentColumn() 
     
    -
    +
     byte[]
     getFamily() 
     
    -
    +
     byte[]
     getQualifier() 
     
    -
    +
     boolean
     hasFilterRow()
     Fitlers that never filter by modifying the returned List of 
    Cells can
      inherit this implementation that does nothing.
     
     
    -
    +
     static DependentColumnFilter
     parseFrom(byte[] pbBytes) 
     
    -
    +
     void
     reset()
     Filters that are purely stateless and do nothing in their 
    reset() methods can inherit
      this null/empty implementation.
     
     
    -
    +
     byte[]
     toByteArray()
     Return length 0 byte array for Filters that don't require 
    special serialization
     
     
    -
    +
     http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String
     toString()
     Return filter's info for debugging and logging 
    purpose.
    @@ -620,7 +626,9 @@ public 
     
     filterKeyValue
    -public Filter.ReturnCode filterKeyValue(Cell c)
    +http://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true";
     title="class or interface in java.lang">@Deprecated
    +public Filter.ReturnCode filterKeyValue(Cell c)
    +Deprecated. 
     Description copied from 
    class: Filter
     A way to filter based on the column family, column 
    qualifier and/or the column value. Return
      code is described below. This allows filters to filter only certain number of 
    columns, then
    @@ -633,15 +641,49 @@ public Filter.reset()
     is called just in case the caller calls
      for the next row.
    - 
    +
      Concrete implementers can signal a failure condition in their code by 
    throwing an
      http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
     title="class or interface in java.io">IOException.
     
    -Specified by:
    +Overrides:
     filterKeyValue in
     class Filter
     Parameters:
     c - the Cell in question
     Returns:
    +code as described below, Filter.ReturnCode.INCLUDE by default
    +See Also:
    +Filter.ReturnCode
    +
    +
    +
    +
    +
    +
    +
    +
    +filterCell
    +public Filter.ReturnCode filterCell(Cell c)
    +Description copied from 
    class: Filter
    +A way to filter based on the column family, column 
    qualifier and/or the column value. Return
    + code is described below. This allows filters to filter only certain number of 
    columns, then
    + terminate without matching ever column.
    +
    + If filterRowKey returns true, filterCell needs to be consistent with it.
    +
    + filterCell can assume that filterRowKey has already been called for the row.
    +
    + If your filter returns ReturnCode.NEXT_ROW, it should return
    + ReturnCode.NEXT_ROW until Filter.reset()
     is called just in case the caller calls
    + for the next row.
    +
    + Concrete implementers can signal a failure condition in their code by 
    throwing an
    + http://docs.oracle.com/javase/8/docs/api/java/i

    [22/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html
    --
    diff --git 
    a/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html 
    b/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html
    index 8b4a6be..b6c5f3f 100644
    --- a/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html
    +++ b/apidocs/src-html/org/apache/hadoop/hbase/rest/client/RemoteHTable.html
    @@ -82,847 +82,871 @@
     074import java.util.Map;
     075import java.util.Set;
     076import java.util.TreeMap;
    -077
    -078/**
    -079 * HTable interface to remote tables 
    accessed via REST gateway
    -080 */
    -081@InterfaceAudience.Public
    -082public class RemoteHTable implements 
    Table {
    -083
    -084  private static final Log LOG = 
    LogFactory.getLog(RemoteHTable.class);
    -085
    -086  final Client client;
    -087  final Configuration conf;
    -088  final byte[] name;
    -089  final int maxRetries;
    -090  final long sleepTime;
    -091
    -092  @SuppressWarnings("rawtypes")
    -093  protected String buildRowSpec(final 
    byte[] row, final Map familyMap,
    -094  final long startTime, final long 
    endTime, final int maxVersions) {
    -095StringBuffer sb = new 
    StringBuffer();
    -096sb.append('/');
    -097sb.append(Bytes.toString(name));
    -098sb.append('/');
    -099sb.append(toURLEncodedBytes(row));
    -100Set families = 
    familyMap.entrySet();
    -101if (families != null) {
    -102  Iterator i = 
    familyMap.entrySet().iterator();
    -103  sb.append('/');
    -104  while (i.hasNext()) {
    -105Map.Entry e = 
    (Map.Entry)i.next();
    -106Collection quals = 
    (Collection)e.getValue();
    -107if (quals == null || 
    quals.isEmpty()) {
    -108  // this is an unqualified 
    family. append the family name and NO ':'
    -109  
    sb.append(toURLEncodedBytes((byte[])e.getKey()));
    -110} else {
    -111  Iterator ii = 
    quals.iterator();
    -112  while (ii.hasNext()) {
    -113
    sb.append(toURLEncodedBytes((byte[])e.getKey()));
    -114sb.append(':');
    -115Object o = ii.next();
    -116// Puts use byte[] but 
    Deletes use KeyValue
    -117if (o instanceof byte[]) {
    -118  
    sb.append(toURLEncodedBytes((byte[])o));
    -119} else if (o instanceof 
    KeyValue) {
    -120  
    sb.append(toURLEncodedBytes(CellUtil.cloneQualifier((KeyValue)o)));
    -121} else {
    -122  throw new 
    RuntimeException("object type not handled");
    -123}
    -124if (ii.hasNext()) {
    -125  sb.append(',');
    -126}
    -127  }
    -128}
    -129if (i.hasNext()) {
    -130  sb.append(',');
    -131}
    -132  }
    -133}
    -134if (startTime >= 0 && 
    endTime != Long.MAX_VALUE) {
    -135  sb.append('/');
    -136  sb.append(startTime);
    -137  if (startTime != endTime) {
    -138sb.append(',');
    -139sb.append(endTime);
    -140  }
    -141} else if (endTime != Long.MAX_VALUE) 
    {
    -142  sb.append('/');
    -143  sb.append(endTime);
    -144}
    -145if (maxVersions > 1) {
    -146  sb.append("?v=");
    -147  sb.append(maxVersions);
    -148}
    -149return sb.toString();
    -150  }
    -151
    -152  protected String 
    buildMultiRowSpec(final byte[][] rows, int maxVersions) {
    -153StringBuilder sb = new 
    StringBuilder();
    -154sb.append('/');
    -155sb.append(Bytes.toString(name));
    -156sb.append("/multiget/");
    -157if (rows == null || rows.length == 0) 
    {
    -158  return sb.toString();
    -159}
    -160sb.append("?");
    -161for(int i=0; i
    

    [26/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/src-html/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.html
    --
    diff --git 
    a/apidocs/src-html/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.html 
    b/apidocs/src-html/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.html
    index d780353..1554d23 100644
    --- a/apidocs/src-html/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.html
    +++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.html
    @@ -132,387 +132,393 @@
     124return false;
     125  }
     126
    -127  @Override
    -128  public ReturnCode filterKeyValue(Cell 
    ignored) {
    -129return currentReturnCode;
    -130  }
    -131
    -132  @Override
    -133  public Cell getNextCellHint(Cell 
    currentKV) {
    -134// skip to the next range's start 
    row
    -135return 
    CellUtil.createFirstOnRow(range.startRow, 0,
    -136(short) range.startRow.length);
    -137  }
    -138
    -139  /**
    -140   * @return The filter serialized using 
    pb
    -141   */
    -142  public byte[] toByteArray() {
    -143
    FilterProtos.MultiRowRangeFilter.Builder builder = 
    FilterProtos.MultiRowRangeFilter
    -144.newBuilder();
    -145for (RowRange range : rangeList) {
    -146  if (range != null) {
    -147FilterProtos.RowRange.Builder 
    rangebuilder = FilterProtos.RowRange.newBuilder();
    -148if (range.startRow != null)
    -149  
    rangebuilder.setStartRow(UnsafeByteOperations.unsafeWrap(range.startRow));
    -150
    rangebuilder.setStartRowInclusive(range.startRowInclusive);
    -151if (range.stopRow != null)
    -152  
    rangebuilder.setStopRow(UnsafeByteOperations.unsafeWrap(range.stopRow));
    -153
    rangebuilder.setStopRowInclusive(range.stopRowInclusive);
    -154
    builder.addRowRangeList(rangebuilder.build());
    -155  }
    -156}
    -157return 
    builder.build().toByteArray();
    -158  }
    -159
    -160  /**
    -161   * @param pbBytes A pb serialized 
    instance
    -162   * @return An instance of 
    MultiRowRangeFilter
    -163   * @throws 
    org.apache.hadoop.hbase.exceptions.DeserializationException
    -164   */
    -165  public static MultiRowRangeFilter 
    parseFrom(final byte[] pbBytes)
    -166  throws DeserializationException {
    -167FilterProtos.MultiRowRangeFilter 
    proto;
    -168try {
    -169  proto = 
    FilterProtos.MultiRowRangeFilter.parseFrom(pbBytes);
    -170} catch 
    (InvalidProtocolBufferException e) {
    -171  throw new 
    DeserializationException(e);
    -172}
    -173int length = 
    proto.getRowRangeListCount();
    -174List 
    rangeProtos = proto.getRowRangeListList();
    -175List rangeList = new 
    ArrayList<>(length);
    -176for (FilterProtos.RowRange rangeProto 
    : rangeProtos) {
    -177  RowRange range = new 
    RowRange(rangeProto.hasStartRow() ? rangeProto.getStartRow()
    -178  .toByteArray() : null, 
    rangeProto.getStartRowInclusive(), rangeProto.hasStopRow() ?
    -179  
    rangeProto.getStopRow().toByteArray() : null, 
    rangeProto.getStopRowInclusive());
    -180  rangeList.add(range);
    -181}
    -182return new 
    MultiRowRangeFilter(rangeList);
    -183  }
    -184
    -185  /**
    -186   * @param o the filter to compare
    -187   * @return true if and only if the 
    fields of the filter that are serialized are equal to the
    -188   * corresponding fields in 
    other. Used for testing.
    -189   */
    -190  boolean areSerializedFieldsEqual(Filter 
    o) {
    -191if (o == this)
    -192  return true;
    -193if (!(o instanceof 
    MultiRowRangeFilter))
    -194  return false;
    -195
    -196MultiRowRangeFilter other = 
    (MultiRowRangeFilter) o;
    -197if (this.rangeList.size() != 
    other.rangeList.size())
    -198  return false;
    -199for (int i = 0; i < 
    rangeList.size(); ++i) {
    -200  RowRange thisRange = 
    this.rangeList.get(i);
    -201  RowRange otherRange = 
    other.rangeList.get(i);
    -202  if 
    (!(Bytes.equals(thisRange.startRow, otherRange.startRow) && 
    Bytes.equals(
    -203  thisRange.stopRow, 
    otherRange.stopRow) && (thisRange.startRowInclusive ==
    -204  otherRange.startRowInclusive) 
    && (thisRange.stopRowInclusive ==
    -205  otherRange.stopRowInclusive))) 
    {
    -206return false;
    -207  }
    -208}
    -209return true;
    -210  }
    -211
    -212  /**
    -213   * calculate the position where the row 
    key in the ranges list.
    -214   *
    -215   * @param rowKey the row key to 
    calculate
    -216   * @return index the position of the 
    row key
    -217   */
    -218  private int getNextRangeIndex(byte[] 
    rowKey) {
    -219RowRange temp = new RowRange(rowKey, 
    true, null, true);
    -220int index = 
    Collections.binarySearch(rangeList, temp);
    -221if (index < 0) {
    -222  int insertionPosition = -index - 
    1;
    -223  // check if the row key in the 
    range before the insertion position
    -224  if (insertionPosition != 0 
    && rangeList.get(insertionPosition - 1).contains(rowKey)) {
    -225return insertionPosition - 1;
    -226  }
    -227  // check if the row key is before 
    the first range
    -228  if (i

    [38/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/org/apache/hadoop/hbase/filter/SkipFilter.html
    --
    diff --git a/apidocs/org/apache/hadoop/hbase/filter/SkipFilter.html 
    b/apidocs/org/apache/hadoop/hbase/filter/SkipFilter.html
    index f6ea339..2ef288d 100644
    --- a/apidocs/org/apache/hadoop/hbase/filter/SkipFilter.html
    +++ b/apidocs/org/apache/hadoop/hbase/filter/SkipFilter.html
    @@ -18,8 +18,8 @@
     catch(err) {
     }
     //-->
    -var methods = 
    {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":9,"i7":10,"i8":10,"i9":10,"i10":10};
    -var tabs = {65535:["t0","All Methods"],1:["t1","Static 
    Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]};
    +var methods = 
    {"i0":10,"i1":42,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":9,"i8":10,"i9":10,"i10":10,"i11":10};
    +var tabs = {65535:["t0","All Methods"],1:["t1","Static 
    Methods"],2:["t2","Instance Methods"],8:["t4","Concrete 
    Methods"],32:["t6","Deprecated Methods"]};
     var altColor = "altColor";
     var rowColor = "rowColor";
     var tableTab = "tableTab";
    @@ -201,72 +201,78 @@ extends org.apache.hadoop.hbase.filter.FilterBase
     
     Method Summary
     
    -All Methods Static Methods Instance Methods Concrete Methods 
    +All Methods Static Methods Instance Methods Concrete Methods Deprecated Methods 
     
     Modifier and Type
     Method and Description
     
     
     Filter.ReturnCode
    -filterKeyValue(Cell v)
    +filterCell(Cell c)
     A way to filter based on the column family, column 
    qualifier and/or the column value.
     
     
     
    +Filter.ReturnCode
    +filterKeyValue(Cell c)
    +Deprecated. 
    +
    +
    +
     boolean
     filterRow()
     Filters that never filter by rows based on previously 
    gathered state from
    - Filter.filterKeyValue(Cell)
     can inherit this implementation that
    + Filter.filterCell(Cell)
     can inherit this implementation that
      never filters a row.
     
     
    -
    +
     boolean
     filterRowKey(Cell cell)
     Filters a row based on the row key.
     
     
    -
    +
     Filter
     getFilter() 
     
    -
    +
     boolean
     hasFilterRow()
     Fitlers that never filter by modifying the returned List of 
    Cells can
      inherit this implementation that does nothing.
     
     
    -
    +
     boolean
     isFamilyEssential(byte[] name)
     By default, we require all scan's column families to be 
    present.
     
     
    -
    +
     static SkipFilter
     parseFrom(byte[] pbBytes) 
     
    -
    +
     void
     reset()
     Filters that are purely stateless and do nothing in their 
    reset() methods can inherit
      this null/empty implementation.
     
     
    -
    +
     byte[]
     toByteArray()
     Return length 0 byte array for Filters that don't require 
    special serialization
     
     
    -
    +
     http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String
     toString()
     Return filter's info for debugging and logging 
    purpose.
     
     
    -
    +
     Cell
     transformCell(Cell v)
     By default no transformation takes place
    @@ -370,7 +376,7 @@ extends org.apache.hadoop.hbase.filter.FilterBase
      throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
     title="class or interface in java.io">IOException
     Description copied from 
    class: Filter
     Filters a row based on the row key. If this returns true, 
    the entire row will be excluded. If
    - false, each KeyValue in the row will be passed to Filter.filterKeyValue(Cell)
     below.
    + false, each KeyValue in the row will be passed to Filter.filterCell(Cell)
     below.
      If Filter.filterAllRemaining()
     returns true, then Filter.filterRowKey(Cell)
     should
      also return true.
     
    @@ -394,8 +400,10 @@ extends org.apache.hadoop.hbase.filter.FilterBase
     
     
     filterKeyValue
    -public Filter.ReturnCode filterKeyValue(Cell v)
    - throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
     title="class or interface in java.io">IOException
    +http://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true";
     title="class or interface in java.lang">@Deprecated
    +public Filter.ReturnCode filterKeyValue(Cell c)
    + throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
     title="class or interface in java.io">IOException
    +Deprecated. 
     Description copied from 
    class: Filter
     A way to filter based on the column family, column 
    qualifier and/or the column value. Return
      code is described below. This allows filters to filter only certain number of 
    columns, then
    @@ -408,14 +416,51 @@ extends org.apache.hadoop.hbase.filter.FilterBase
      If your filter returns ReturnCode.NEXT_ROW, it should return
      ReturnCode.NEXT_ROW until Filter.reset()
     is called just in case the caller calls
      for the next row.
    - 
    +
      Concrete implementers can signal a failure condition in their code by 
    throwing an
      http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
     title="class or interface in java.io">IOException.
     
    -Specified by:
    +Overrides:
     filterKeyValue in
     class Filter
     Parameters:
    -v - the Cell in questi

    [04/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/devapidocs/org/apache/hadoop/hbase/filter/FilterListBase.html
    --
    diff --git a/devapidocs/org/apache/hadoop/hbase/filter/FilterListBase.html 
    b/devapidocs/org/apache/hadoop/hbase/filter/FilterListBase.html
    index dfa465f..9e0a2b4 100644
    --- a/devapidocs/org/apache/hadoop/hbase/filter/FilterListBase.html
    +++ b/devapidocs/org/apache/hadoop/hbase/filter/FilterListBase.html
    @@ -18,7 +18,7 @@
     catch(err) {
     }
     //-->
    -var methods = 
    {"i0":6,"i1":9,"i2":10,"i3":10,"i4":10,"i5":6,"i6":10,"i7":10,"i8":6,"i9":10,"i10":10,"i11":9,"i12":10,"i13":10,"i14":10,"i15":10};
    +var methods = 
    {"i0":6,"i1":9,"i2":10,"i3":10,"i4":10,"i5":10,"i6":6,"i7":10,"i8":10,"i9":6,"i10":10,"i11":10,"i12":9,"i13":10,"i14":10,"i15":10,"i16":10};
     var tabs = {65535:["t0","All Methods"],1:["t1","Static 
    Methods"],2:["t2","Instance Methods"],4:["t3","Abstract 
    Methods"],8:["t4","Concrete Methods"]};
     var altColor = "altColor";
     var rowColor = "rowColor";
    @@ -178,7 +178,7 @@ extends 
     protected Cell
     transformedCell
    -When filtering a given Cell in filterKeyValue(Cell),
     this stores the transformed Cell
    +When filtering a given Cell in filterCell(Cell),
     this stores the transformed Cell
      to be returned by transformCell(Cell).
     
     
    @@ -237,71 +237,77 @@ extends 
     
     Filter.ReturnCode
    -filterKeyValue(Cell c)
    +filterCell(Cell c)
     A way to filter based on the column family, column 
    qualifier and/or the column value.
     
     
     
    +Filter.ReturnCode
    +filterKeyValue(Cell c)
    +A way to filter based on the column family, column 
    qualifier and/or the column value.
    +
    +
    +
     void
     filterRowCells(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
     title="class or interface in java.util">List cells)
     Filters that never filter by modifying the returned List of 
    Cells can inherit this
      implementation that does nothing.
     
     
    -
    +
     protected abstract http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String
     formatLogFilters(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true";
     title="class or interface in java.util">List logFilters) 
     
    -
    +
     http://docs.oracle.com/javase/8/docs/api/java/util/ArrayList.html?is-external=true";
     title="class or interface in java.util">ArrayList
     getFilters() 
     
    -
    +
     boolean
     hasFilterRow()
     Fitlers that never filter by modifying the returned List of 
    Cells can
      inherit this implementation that does nothing.
     
     
    -
    +
     (package private) abstract Filter.ReturnCode
    -internalFilterKeyValue(Cell c,
    -  Cell transformedCell)
    -Internal implementation of filterKeyValue(Cell)
    +internalFilterCell(Cell c,
    +  Cell transformedCell)
    +Internal implementation of filterCell(Cell)
     
     
    -
    +
     boolean
     isEmpty() 
     
    -
    +
     boolean
     isFamilyEssential(byte[] name)
     By default, we require all scan's column families to be 
    present.
     
     
    -
    +
     protected static boolean
     isInReturnCodes(Filter.ReturnCode testRC,
    Filter.ReturnCode... returnCodes) 
     
    -
    +
     void
     setReversed(boolean reversed)
     alter the reversed scan flag
     
     
    -
    +
     int
     size() 
     
    -
    +
     http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String
     toString()
     Return filter's info for debugging and logging 
    purpose.
     
     
    -
    +
     Cell
     transformCell(Cell c)
     By default no transformation takes place
    @@ -384,7 +390,7 @@ extends 
     transformedCell
     protected Cell transformedCell
    -When filtering a given Cell in filterKeyValue(Cell),
     this stores the transformed Cell
    +When filtering a given Cell in filterCell(Cell),
     this stores the transformed Cell
      to be returned by transformCell(Cell).
     Individual filters transformation are applied
      only when the filter includes the Cell. Transformations are composed in the 
    order specified by
      filters.
    @@ -514,16 +520,16 @@ extends 
     
     
    -
    +
     
     
     
     
    -internalFilterKeyValue
    -abstract Filter.ReturnCode internalFilterKeyValue(Cell c,
    -  Cell transformedCell)
    -   throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
     title="class or interface in java.io">IOException
    -Internal implementation of filterKeyValue(Cell)
    +internalFilterCell
    +abstract Filter.ReturnCode internalFilterCell(Cell c,
    +  Cell transformedCell)
    +   throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
     title="class or interface in java.io">IOException
    +Internal implementation of filterCell(Cell)
     
     Parameters:
     c - The cell in question.
    @@ -533,7 +539,7 @@ extends Throws:
     http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true";
     title="class or interface in java

    [37/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/org/apache/hadoop/hbase/filter/class-use/Filter.ReturnCode.html
    --
    diff --git 
    a/apidocs/org/apache/hadoop/hbase/filter/class-use/Filter.ReturnCode.html 
    b/apidocs/org/apache/hadoop/hbase/filter/class-use/Filter.ReturnCode.html
    index 7cf297c..30331ac 100644
    --- a/apidocs/org/apache/hadoop/hbase/filter/class-use/Filter.ReturnCode.html
    +++ b/apidocs/org/apache/hadoop/hbase/filter/class-use/Filter.ReturnCode.html
    @@ -107,123 +107,276 @@
     
     
     Filter.ReturnCode
    -MultipleColumnPrefixFilter.filterColumn(Cell cell) 
    +FilterList.filterCell(Cell c) 
     
     
     Filter.ReturnCode
    -ColumnPrefixFilter.filterColumn(Cell cell) 
    +WhileMatchFilter.filterCell(Cell c) 
    +
    +
    +Filter.ReturnCode
    +PageFilter.filterCell(Cell ignored) 
    +
    +
    +Filter.ReturnCode
    +MultipleColumnPrefixFilter.filterCell(Cell c) 
    +
    +
    +Filter.ReturnCode
    +InclusiveStopFilter.filterCell(Cell c) 
    +
    +
    +Filter.ReturnCode
    +KeyOnlyFilter.filterCell(Cell ignored) 
    +
    +
    +Filter.ReturnCode
    +RowFilter.filterCell(Cell v) 
    +
    +
    +Filter.ReturnCode
    +ColumnRangeFilter.filterCell(Cell c) 
    +
    +
    +Filter.ReturnCode
    +FamilyFilter.filterCell(Cell c) 
    +
    +
    +Filter.ReturnCode
    +RandomRowFilter.filterCell(Cell c) 
    +
    +
    +Filter.ReturnCode
    +FirstKeyValueMatchingQualifiersFilter.filterCell(Cell c)
    +Deprecated. 
    + 
    +
    +
    +Filter.ReturnCode
    +SkipFilter.filterCell(Cell c) 
     
     
     Filter.ReturnCode
    -FilterList.filterKeyValue(Cell c) 
    +DependentColumnFilter.filterCell(Cell c) 
     
     
     Filter.ReturnCode
    -WhileMatchFilter.filterKeyValue(Cell v) 
    +Filter.filterCell(Cell c)
    +A way to filter based on the column family, column 
    qualifier and/or the column value.
    +
    +
    +
    +Filter.ReturnCode
    +ColumnPaginationFilter.filterCell(Cell c) 
    +
    +
    +Filter.ReturnCode
    +ValueFilter.filterCell(Cell c) 
    +
    +
    +Filter.ReturnCode
    +ColumnCountGetFilter.filterCell(Cell c) 
    +
    +
    +Filter.ReturnCode
    +QualifierFilter.filterCell(Cell c) 
    +
    +
    +Filter.ReturnCode
    +PrefixFilter.filterCell(Cell c) 
    +
    +
    +Filter.ReturnCode
    +FuzzyRowFilter.filterCell(Cell c) 
     
     
     Filter.ReturnCode
    -PageFilter.filterKeyValue(Cell ignored) 
    +TimestampsFilter.filterCell(Cell c) 
     
     
     Filter.ReturnCode
    -MultipleColumnPrefixFilter.filterKeyValue(Cell kv) 
    +ColumnPrefixFilter.filterCell(Cell cell) 
     
     
     Filter.ReturnCode
    -InclusiveStopFilter.filterKeyValue(Cell v) 
    +MultiRowRangeFilter.filterCell(Cell ignored) 
     
     
     Filter.ReturnCode
    -KeyOnlyFilter.filterKeyValue(Cell ignored) 
    +SingleColumnValueFilter.filterCell(Cell c) 
     
     
     Filter.ReturnCode
    -RowFilter.filterKeyValue(Cell v) 
    +FirstKeyOnlyFilter.filterCell(Cell c) 
     
     
     Filter.ReturnCode
    -ColumnRangeFilter.filterKeyValue(Cell kv) 
    +MultipleColumnPrefixFilter.filterColumn(Cell cell) 
     
     
     Filter.ReturnCode
    -FamilyFilter.filterKeyValue(Cell v) 
    +ColumnPrefixFilter.filterColumn(Cell cell) 
     
     
     Filter.ReturnCode
    -RandomRowFilter.filterKeyValue(Cell v) 
    +FilterList.filterKeyValue(Cell c)
    +Deprecated. 
    +
     
     
     Filter.ReturnCode
    -FirstKeyValueMatchingQualifiersFilter.filterKeyValue(Cell v)
    +WhileMatchFilter.filterKeyValue(Cell c)
     Deprecated. 
    - 
    +
     
     
     Filter.ReturnCode
    -SkipFilter.filterKeyValue(Cell v) 
    +PageFilter.filterKeyValue(Cell c)
    +Deprecated. 
    +
     
     
     Filter.ReturnCode
    -DependentColumnFilter.filterKeyValue(Cell c) 
    +MultipleColumnPrefixFilter.filterKeyValue(Cell c)
    +Deprecated. 
    +
     
     
    -abstract Filter.ReturnCode
    -Filter.filterKeyValue(Cell v)
    -A way to filter based on the column family, column 
    qualifier and/or the column value.
    +Filter.ReturnCode
    +InclusiveStopFilter.filterKeyValue(Cell c)
    +Deprecated. 
    +
    +
    +
    +Filter.ReturnCode
    +KeyOnlyFilter.filterKeyValue(Cell ignored)
    +Deprecated. 
    +
    +
    +
    +Filter.ReturnCode
    +RowFilter.filterKeyValue(Cell c)
    +Deprecated. 
    +
    +
    +
    +Filter.ReturnCode
    +ColumnRangeFilter.filterKeyValue(Cell c)
    +Deprecated. 
    +
    +
    +
    +Filter.ReturnCode
    +FamilyFilter.filterKeyValue(Cell c)
    +Deprecated. 
    +
    +
    +
    +Filter.ReturnCode
    +RandomRowFilter.filterKeyValue(Cell c)
    +Deprecated. 
    +
    +
    +
    +Filter.ReturnCode
    +FirstKeyValueMatchingQualifiersFilter.filterKeyValue(Cell c)
    +Deprecated. 
    +
    +
    +
    +Filter.ReturnCode
    +SkipFilter.filterKeyValue(Cell c)
    +Deprecated. 
    +
    +
    +
    +Filter.ReturnCode
    +DependentColumnFilter.filterKeyValue(Cell c)
    +Deprecated. 
     
     
     
     Filter.ReturnCode
    -ColumnPaginationFilter.filterKeyValue(Cell v) 
    +Filter.filterKeyValue(Cell c)
    +Deprecated. 
    +As of release 2.0.0, this 
    will be removed in HBase 3.0.0.
    + Instead use filterCell(Cell)
    +
    +
     
     
     Filter.ReturnCode
    -ValueFilter.filterKeyValue(Cell v) 
    +ColumnPaginationFilter.filterKeyValue(Cell c)
    +Deprecated. 
    +
     
     
     Filter.ReturnCode
    -ColumnCountGetFilter.filterKeyValue(Cell v) 
    +ValueFilter.filterKeyValue(Cell c)
    +Deprecated. 
    +
     
     
     Filter.ReturnCode
    -QualifierFilter.filterKeyValue(Cell v) 
    +ColumnCountGetFilter.filterKeyValue(Cell c)
    +Deprecated. 
    +
     
     
     Filter.ReturnCode
    -PrefixFilter.filterKeyValue(Cell v) 
    +QualifierFilter.filterKeyValue(Cell c)
    +Deprecated. 
    +
     
     
     Filter.ReturnCode
    -FuzzyRowFilter.fi

    [28/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/src-html/org/apache/hadoop/hbase/filter/InclusiveStopFilter.html
    --
    diff --git 
    a/apidocs/src-html/org/apache/hadoop/hbase/filter/InclusiveStopFilter.html 
    b/apidocs/src-html/org/apache/hadoop/hbase/filter/InclusiveStopFilter.html
    index 23e0f62..ced79d7 100644
    --- a/apidocs/src-html/org/apache/hadoop/hbase/filter/InclusiveStopFilter.html
    +++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/InclusiveStopFilter.html
    @@ -59,77 +59,83 @@
     051return this.stopRowKey;
     052  }
     053
    -054  @Override
    -055  public ReturnCode filterKeyValue(Cell 
    v) {
    -056if (done) return 
    ReturnCode.NEXT_ROW;
    -057return ReturnCode.INCLUDE;
    +054  @Deprecated
    +055  @Override
    +056  public ReturnCode filterKeyValue(final 
    Cell c) {
    +057return filterCell(c);
     058  }
     059
    -060  public boolean filterRowKey(Cell 
    firstRowCell) {
    -061// if stopRowKey is <= buffer, 
    then true, filter row.
    -062if (filterAllRemaining()) return 
    true;
    -063int cmp = 
    CellComparatorImpl.COMPARATOR.compareRows(firstRowCell, stopRowKey, 0, 
    stopRowKey.length);
    -064done = reversed ? cmp < 0 : cmp 
    > 0;
    -065return done;
    -066  }
    -067
    -068  public boolean filterAllRemaining() {
    -069return done;
    -070  }
    -071
    -072  public static Filter 
    createFilterFromArguments (ArrayList filterArguments) {
    -073
    Preconditions.checkArgument(filterArguments.size() == 1,
    -074"Expected 
    1 but got: %s", filterArguments.size());
    -075byte [] stopRowKey = 
    ParseFilter.removeQuotesFromByteArray(filterArguments.get(0));
    -076return new 
    InclusiveStopFilter(stopRowKey);
    -077  }
    -078
    -079  /**
    -080   * @return The filter serialized using 
    pb
    -081   */
    -082  public byte [] toByteArray() {
    -083
    FilterProtos.InclusiveStopFilter.Builder builder =
    -084  
    FilterProtos.InclusiveStopFilter.newBuilder();
    -085if (this.stopRowKey != null) 
    builder.setStopRowKey(
    -086
    UnsafeByteOperations.unsafeWrap(this.stopRowKey));
    -087return 
    builder.build().toByteArray();
    -088  }
    -089
    -090  /**
    -091   * @param pbBytes A pb serialized 
    {@link InclusiveStopFilter} instance
    -092   * @return An instance of {@link 
    InclusiveStopFilter} made from bytes
    -093   * @throws DeserializationException
    -094   * @see #toByteArray
    -095   */
    -096  public static InclusiveStopFilter 
    parseFrom(final byte [] pbBytes)
    -097  throws DeserializationException {
    -098FilterProtos.InclusiveStopFilter 
    proto;
    -099try {
    -100  proto = 
    FilterProtos.InclusiveStopFilter.parseFrom(pbBytes);
    -101} catch 
    (InvalidProtocolBufferException e) {
    -102  throw new 
    DeserializationException(e);
    -103}
    -104return new 
    InclusiveStopFilter(proto.hasStopRowKey()?proto.getStopRowKey().toByteArray():null);
    -105  }
    -106
    -107  /**
    -108   * @param other
    -109   * @return true if and only if the 
    fields of the filter that are serialized
    -110   * are equal to the corresponding 
    fields in other.  Used for testing.
    -111   */
    -112  boolean areSerializedFieldsEqual(Filter 
    o) {
    -113if (o == this) return true;
    -114if (!(o instanceof 
    InclusiveStopFilter)) return false;
    -115
    -116InclusiveStopFilter other = 
    (InclusiveStopFilter)o;
    -117return 
    Bytes.equals(this.getStopRowKey(), other.getStopRowKey());
    -118  }
    -119
    -120  @Override
    -121  public String toString() {
    -122return 
    this.getClass().getSimpleName() + " " + 
    Bytes.toStringBinary(this.stopRowKey);
    -123  }
    -124}
    +060  @Override
    +061  public ReturnCode filterCell(final Cell 
    c) {
    +062if (done) return 
    ReturnCode.NEXT_ROW;
    +063return ReturnCode.INCLUDE;
    +064  }
    +065
    +066  public boolean filterRowKey(Cell 
    firstRowCell) {
    +067// if stopRowKey is <= buffer, 
    then true, filter row.
    +068if (filterAllRemaining()) return 
    true;
    +069int cmp = 
    CellComparatorImpl.COMPARATOR.compareRows(firstRowCell, stopRowKey, 0, 
    stopRowKey.length);
    +070done = reversed ? cmp < 0 : cmp 
    > 0;
    +071return done;
    +072  }
    +073
    +074  public boolean filterAllRemaining() {
    +075return done;
    +076  }
    +077
    +078  public static Filter 
    createFilterFromArguments (ArrayList filterArguments) {
    +079
    Preconditions.checkArgument(filterArguments.size() == 1,
    +080"Expected 
    1 but got: %s", filterArguments.size());
    +081byte [] stopRowKey = 
    ParseFilter.removeQuotesFromByteArray(filterArguments.get(0));
    +082return new 
    InclusiveStopFilter(stopRowKey);
    +083  }
    +084
    +085  /**
    +086   * @return The filter serialized using 
    pb
    +087   */
    +088  public byte [] toByteArray() {
    +089
    FilterProtos.InclusiveStopFilter.Builder builder =
    +090  
    FilterProtos.InclusiveStopFilter.newBuilder();
    +091if (this.stopRowKey != null) 
    builder.setStopRowKey(
    +092
    UnsafeByteOperations.unsafeWrap(this.stopRowKey));
    +093return 
    builder.build().toByteArray();
    +094  }
    +095
    +096  /**
    +097   * @para

    [15/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
    --
    diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html 
    b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
    index 3d1869a..07abd49 100644
    --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
    +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html
    @@ -5246,11 +5246,10 @@ service.
     
     
     default void
    -MasterObserver.postRequestLock(ObserverContext ctx,
    +MasterObserver.postRequestLock(ObserverContext ctx,
    http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String namespace,
    TableName tableName,
    RegionInfo[] regionInfos,
    -   LockType type,
    http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String description)
     Called after new LockProcedure is queued.
     
    @@ -5332,13 +5331,21 @@ service.
     
     
     default void
    +MasterObserver.preLockHeartbeat(ObserverContext ctx,
    +TableName tn,
    +http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String description)
    +Called before heartbeat to a lock.
    +
    +
    +
    +default void
     MasterObserver.preModifyTable(ObserverContext ctx,
       TableName tableName,
       TableDescriptor htd)
     Called prior to modifying a table's properties.
     
     
    -
    +
     default void
     MasterObserver.preModifyTableAction(ObserverContext ctx,
     TableName tableName,
    @@ -5346,18 +5353,17 @@ service.
     Called prior to modifying a table's properties.
     
     
    -
    +
     default void
    -MasterObserver.preRequestLock(ObserverContext ctx,
    +MasterObserver.preRequestLock(ObserverContext ctx,
       http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String namespace,
       TableName tableName,
       RegionInfo[] regionInfos,
    -  LockType type,
       http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String description)
     Called before new LockProcedure is queued.
     
     
    -
    +
     default void
     MasterObserver.preSetTableQuota(ObserverContext ctx,
     TableName tableName,
    @@ -5365,7 +5371,7 @@ service.
     Called before the quota for the table is stored.
     
     
    -
    +
     default void
     MasterObserver.preSetUserQuota(ObserverContext ctx,
    http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in java.lang">String userName,
    @@ -5374,7 +5380,7 @@ service.
     Called before the quota for the user on the specified table 
    is stored.
     
     
    -
    +
     default void
     MasterObserver.preSplitRegion(ObserverContext c,
       TableName tableName,
    @@ -5382,7 +5388,7 @@ service.
     Called before the split region procedure is called.
     
     
    -
    +
     default void
     MasterObserver.preSplitRegionAction(ObserverContext c,
     TableName tableName,
    @@ -5390,14 +5396,14 @@ service.
     Called before the region is split.
     
     
    -
    +
     default void
     MasterObserver.preTableFlush(ObserverContext ctx,
      TableName tableName)
     Called before the table memstore is flushed to disk.
     
     
    -
    +
     default void
     MasterObserver.preTruncateTable(ObserverContext ctx,
     TableName tableName)
    @@ -5405,7 +5411,7 @@ service.
      table.
     
     
    -
    +
     default void
     MasterObserver.preTruncateTableAction(ObserverContext ctx,
       TableName tableName)
    @@ -5413,7 +5419,7 @@ service.
      table.
     
     
    -
    +
     static http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true";
     title="class or interface in java.util">Map
     Export.run(org.apache.hadoop.conf.Configuration conf,
    TableName tableName,
    @@ -9640,64 +9646,69 @@ service.
     
     
     void
    +AccessController.preLockHeartbeat(ObserverContext ctx,
    +TableName tableName,
    +http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true";
     title="class or interface in 
    java.lang">String description) 
    +
    +
    +void
     AccessController.preModifyTable(ObserverContext c,
       TableName tableName,
       TableDescriptor htd) 
     
    -
    +
     

    [14/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/devapidocs/org/apache/hadoop/hbase/client/AsyncTableBase.html
    --
    diff --git a/devapidocs/org/apache/hadoop/hbase/client/AsyncTableBase.html 
    b/devapidocs/org/apache/hadoop/hbase/client/AsyncTableBase.html
    index 289e895..b45f132 100644
    --- a/devapidocs/org/apache/hadoop/hbase/client/AsyncTableBase.html
    +++ b/devapidocs/org/apache/hadoop/hbase/client/AsyncTableBase.html
    @@ -380,10 +380,14 @@ public interface 
     
     getRpcTimeout
    -long getRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
    +long getRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
     Get timeout of each rpc request in this Table instance. It 
    will be overridden by a more
      specific rpc timeout config such as readRpcTimeout or writeRpcTimeout.
     
    +Parameters:
    +unit - the unit of time the timeout to be represented in
    +Returns:
    +rpc timeout in the specified time unit
     See Also:
     getReadRpcTimeout(TimeUnit),
     
     getWriteRpcTimeout(TimeUnit)
    @@ -396,8 +400,14 @@ public interface 
     
     getReadRpcTimeout
    -long getReadRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
    +long getReadRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
     Get timeout of each rpc read request in this Table 
    instance.
    +
    +Parameters:
    +unit - the unit of time the timeout to be represented in
    +Returns:
    +read rpc timeout in the specified time unit
    +
     
     
     
    @@ -406,8 +416,14 @@ public interface 
     
     getWriteRpcTimeout
    -long getWriteRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
    +long getWriteRpcTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
     Get timeout of each rpc write request in this Table 
    instance.
    +
    +Parameters:
    +unit - the unit of time the timeout to be represented in
    +Returns:
    +write rpc timeout in the specified time unit
    +
     
     
     
    @@ -416,8 +432,14 @@ public interface 
     
     getOperationTimeout
    -long getOperationTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
    +long getOperationTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
     Get timeout of each operation in Table instance.
    +
    +Parameters:
    +unit - the unit of time the timeout to be represented in
    +Returns:
    +operation rpc timeout in the specified time unit
    +
     
     
     
    @@ -426,9 +448,15 @@ public interface 
     
     getScanTimeout
    -long getScanTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
    +long getScanTimeout(http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/TimeUnit.html?is-external=true";
     title="class or interface in 
    java.util.concurrent">TimeUnit unit)
     Get the timeout of a single operation in a scan. It works 
    like operation timeout for other
      operations.
    +
    +Parameters:
    +unit - the unit of time the timeout to be represented in
    +Returns:
    +scan rpc timeout in the specified time unit
    +
     
     
     
    @@ -437,7 +465,7 @@ public interface 
     
     exists
    -default http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
     title="class or interface in java.util.concurrent">CompletableFutureBoolean> exists(Get get)
    +default http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
     title="class or interface in java.util.concurrent">CompletableFutureBoolean> exists(Get get)
     Test for the existence of columns in the table, as 
    specified by the Get.
      
      This will return true if the Get matches one or more keys, false if not.
    @@ -456,7 +484,7 @@ public interface 
     
     get
    -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true";
     title="class or interface in java.util.concurrent">CompletableFuture get(Get get)
    +http://docs.oracle.com/javase/8/docs/

    [32/51] [partial] hbase-site git commit: Published site at .

    http://git-wip-us.apache.org/repos/asf/hbase-site/blob/00c22388/apidocs/src-html/org/apache/hadoop/hbase/filter/DependentColumnFilter.html
    --
    diff --git 
    a/apidocs/src-html/org/apache/hadoop/hbase/filter/DependentColumnFilter.html 
    b/apidocs/src-html/org/apache/hadoop/hbase/filter/DependentColumnFilter.html
    index 7cc3e0b..a012923 100644
    --- a/apidocs/src-html/org/apache/hadoop/hbase/filter/DependentColumnFilter.html
    +++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/DependentColumnFilter.html
    @@ -163,155 +163,161 @@
     155return false;
     156  }
     157
    -158  @Override
    -159  public ReturnCode filterKeyValue(Cell 
    c) {
    -160// Check if the column and qualifier 
    match
    -161if (!CellUtil.matchingColumn(c, 
    this.columnFamily, this.columnQualifier)) {
    -162// include non-matches for the 
    time being, they'll be discarded afterwards
    -163return ReturnCode.INCLUDE;
    -164}
    -165// If it doesn't pass the op, skip 
    it
    -166if (comparator != null
    -167&& 
    compareValue(getCompareOperator(), comparator, c))
    -168  return ReturnCode.SKIP;
    -169  
    -170stampSet.add(c.getTimestamp());
    -171if(dropDependentColumn) {
    -172  return ReturnCode.SKIP;
    -173}
    -174return ReturnCode.INCLUDE;
    -175  }
    -176
    -177  @Override
    -178  public void 
    filterRowCells(List kvs) {
    -179kvs.removeIf(kv -> 
    !stampSet.contains(kv.getTimestamp()));
    -180  }
    -181
    -182  @Override
    -183  public boolean hasFilterRow() {
    -184return true;
    -185  }
    -186  
    -187  @Override
    -188  public boolean filterRow() {
    -189return false;
    -190  }
    -191
    -192  @Override
    -193  public boolean filterRowKey(byte[] 
    buffer, int offset, int length) {
    -194return false;
    -195  }
    -196  @Override
    -197  public void reset() {
    -198stampSet.clear();
    -199  }
    -200
    -201  public static Filter 
    createFilterFromArguments(ArrayList filterArguments) {
    -202
    Preconditions.checkArgument(filterArguments.size() == 2 ||
    -203
    filterArguments.size() == 3 ||
    -204
    filterArguments.size() == 5,
    -205"Expected 
    2, 3 or 5 but got: %s", filterArguments.size());
    -206if (filterArguments.size() == 2) {
    -207  byte [] family = 
    ParseFilter.removeQuotesFromByteArray(filterArguments.get(0));
    -208  byte [] qualifier = 
    ParseFilter.removeQuotesFromByteArray(filterArguments.get(1));
    -209  return new 
    DependentColumnFilter(family, qualifier);
    -210
    -211} else if (filterArguments.size() == 
    3) {
    -212  byte [] family = 
    ParseFilter.removeQuotesFromByteArray(filterArguments.get(0));
    -213  byte [] qualifier = 
    ParseFilter.removeQuotesFromByteArray(filterArguments.get(1));
    -214  boolean dropDependentColumn = 
    ParseFilter.convertByteArrayToBoolean(filterArguments.get(2));
    -215  return new 
    DependentColumnFilter(family, qualifier, dropDependentColumn);
    +158  @Deprecated
    +159  @Override
    +160  public ReturnCode filterKeyValue(final 
    Cell c) {
    +161return filterCell(c);
    +162  }
    +163
    +164  @Override
    +165  public ReturnCode filterCell(final Cell 
    c) {
    +166// Check if the column and qualifier 
    match
    +167if (!CellUtil.matchingColumn(c, 
    this.columnFamily, this.columnQualifier)) {
    +168// include non-matches for the 
    time being, they'll be discarded afterwards
    +169return ReturnCode.INCLUDE;
    +170}
    +171// If it doesn't pass the op, skip 
    it
    +172if (comparator != null
    +173&& 
    compareValue(getCompareOperator(), comparator, c))
    +174  return ReturnCode.SKIP;
    +175  
    +176stampSet.add(c.getTimestamp());
    +177if(dropDependentColumn) {
    +178  return ReturnCode.SKIP;
    +179}
    +180return ReturnCode.INCLUDE;
    +181  }
    +182
    +183  @Override
    +184  public void 
    filterRowCells(List kvs) {
    +185kvs.removeIf(kv -> 
    !stampSet.contains(kv.getTimestamp()));
    +186  }
    +187
    +188  @Override
    +189  public boolean hasFilterRow() {
    +190return true;
    +191  }
    +192  
    +193  @Override
    +194  public boolean filterRow() {
    +195return false;
    +196  }
    +197
    +198  @Override
    +199  public boolean filterRowKey(byte[] 
    buffer, int offset, int length) {
    +200return false;
    +201  }
    +202  @Override
    +203  public void reset() {
    +204stampSet.clear();
    +205  }
    +206
    +207  public static Filter 
    createFilterFromArguments(ArrayList filterArguments) {
    +208
    Preconditions.checkArgument(filterArguments.size() == 2 ||
    +209
    filterArguments.size() == 3 ||
    +210
    filterArguments.size() == 5,
    +211"Expected 
    2, 3 or 5 but got: %s", filterArguments.size());
    +212if (filterArguments.size() == 2) {
    +213  byte [] family = 
    ParseFilter.removeQuotesFromByteArray(filterArguments.get(0));
    +214  byte [] qualifier = 
    ParseFilter.removeQuotesFromByteArray(filterArguments.get(1));
    +215  return new 
    DependentColumnFilter(family, qu

    hbase git commit: HBASE-19094 NPE in RSGroupStartupWorker.waitForGroupTableOnline during master startup

    Repository: hbase
    Updated Branches:
      refs/heads/branch-1.4 50b7037e7 -> ac0383830
    
    
    HBASE-19094 NPE in RSGroupStartupWorker.waitForGroupTableOnline during master 
    startup
    
    
    Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
    Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/ac038383
    Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/ac038383
    Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/ac038383
    
    Branch: refs/heads/branch-1.4
    Commit: ac03838303dc17e3e580cfdb9eef18491b45ada1
    Parents: 50b7037
    Author: Abhishek Singh Chouhan 
    Authored: Thu Oct 26 20:17:01 2017 +0530
    Committer: Abhishek Singh Chouhan 
    Committed: Fri Oct 27 17:57:11 2017 +0530
    
    --
     .../org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java  | 4 +++-
     1 file changed, 3 insertions(+), 1 deletion(-)
    --
    
    
    http://git-wip-us.apache.org/repos/asf/hbase/blob/ac038383/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
    --
    diff --git 
    a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
     
    b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
    index eec03ce..80eaefb 100644
    --- 
    a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
    +++ 
    b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
    @@ -640,7 +640,9 @@ public class RSGroupInfoManagerImpl implements 
    RSGroupInfoManager, ServerListene
     if(cell != null) {
       sn = 
    ServerName.parseVersionedServerName(CellUtil.cloneValue(cell));
     }
    -if (tsm.isTableState(TableName.NAMESPACE_TABLE_NAME,
    +if (sn == null) {
    +  nsFound.set(false);
    +} else if (tsm.isTableState(TableName.NAMESPACE_TABLE_NAME,
     ZooKeeperProtos.Table.State.ENABLED)) {
       try {
     ClientProtos.ClientService.BlockingInterface rs = 
    conn.getClient(sn);
    
    
    

    hbase git commit: HBASE-19094 NPE in RSGroupStartupWorker.waitForGroupTableOnline during master startup

    Repository: hbase
    Updated Branches:
      refs/heads/branch-1 3c6261237 -> 0697db220
    
    
    HBASE-19094 NPE in RSGroupStartupWorker.waitForGroupTableOnline during master 
    startup
    
    
    Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
    Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/0697db22
    Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/0697db22
    Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/0697db22
    
    Branch: refs/heads/branch-1
    Commit: 0697db2207da79ecc7774bbaf43007dbc8e6815b
    Parents: 3c62612
    Author: Abhishek Singh Chouhan 
    Authored: Thu Oct 26 20:17:01 2017 +0530
    Committer: Abhishek Singh Chouhan 
    Committed: Fri Oct 27 17:56:01 2017 +0530
    
    --
     .../org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java  | 4 +++-
     1 file changed, 3 insertions(+), 1 deletion(-)
    --
    
    
    http://git-wip-us.apache.org/repos/asf/hbase/blob/0697db22/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
    --
    diff --git 
    a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
     
    b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
    index eec03ce..80eaefb 100644
    --- 
    a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
    +++ 
    b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
    @@ -640,7 +640,9 @@ public class RSGroupInfoManagerImpl implements 
    RSGroupInfoManager, ServerListene
     if(cell != null) {
       sn = 
    ServerName.parseVersionedServerName(CellUtil.cloneValue(cell));
     }
    -if (tsm.isTableState(TableName.NAMESPACE_TABLE_NAME,
    +if (sn == null) {
    +  nsFound.set(false);
    +} else if (tsm.isTableState(TableName.NAMESPACE_TABLE_NAME,
     ZooKeeperProtos.Table.State.ENABLED)) {
       try {
     ClientProtos.ClientService.BlockingInterface rs = 
    conn.getClient(sn);
    
    
    

    hbase git commit: HBASE-19094 NPE in RSGroupStartupWorker.waitForGroupTableOnline during master startup

    Repository: hbase
    Updated Branches:
      refs/heads/branch-2 8e6d116ae -> f981de5bb
    
    
    HBASE-19094 NPE in RSGroupStartupWorker.waitForGroupTableOnline during master 
    startup
    
    
    Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
    Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/f981de5b
    Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/f981de5b
    Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/f981de5b
    
    Branch: refs/heads/branch-2
    Commit: f981de5bb9397c133922f691a1df8e2c18b42ac2
    Parents: 8e6d116
    Author: Abhishek Singh Chouhan 
    Authored: Thu Oct 26 19:40:06 2017 +0530
    Committer: Abhishek Singh Chouhan 
    Committed: Fri Oct 27 17:52:06 2017 +0530
    
    --
     .../org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java  | 4 +++-
     1 file changed, 3 insertions(+), 1 deletion(-)
    --
    
    
    http://git-wip-us.apache.org/repos/asf/hbase/blob/f981de5b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
    --
    diff --git 
    a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
     
    b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
    index e116f58..9520f5f 100644
    --- 
    a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
    +++ 
    b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
    @@ -691,7 +691,9 @@ class RSGroupInfoManagerImpl implements RSGroupInfoManager {
     if(cell != null) {
       sn = 
    ServerName.parseVersionedServerName(CellUtil.cloneValue(cell));
     }
    -if (tsm.isTableState(TableName.NAMESPACE_TABLE_NAME,
    +if (sn == null) {
    +  nsFound.set(false);
    +} else if (tsm.isTableState(TableName.NAMESPACE_TABLE_NAME,
     TableState.State.ENABLED)) {
       try {
     ClientProtos.ClientService.BlockingInterface rs = 
    conn.getClient(sn);
    
    
    

    hbase git commit: HBASE-19094 NPE in RSGroupStartupWorker.waitForGroupTableOnline during master startup

    Repository: hbase
    Updated Branches:
      refs/heads/master 15b32460f -> 4dee4a854
    
    
    HBASE-19094 NPE in RSGroupStartupWorker.waitForGroupTableOnline during master 
    startup
    
    
    Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
    Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/4dee4a85
    Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/4dee4a85
    Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/4dee4a85
    
    Branch: refs/heads/master
    Commit: 4dee4a854fece6516ba5006251898082c8fc161a
    Parents: 15b3246
    Author: Abhishek Singh Chouhan 
    Authored: Thu Oct 26 19:40:06 2017 +0530
    Committer: Abhishek Singh Chouhan 
    Committed: Fri Oct 27 17:50:38 2017 +0530
    
    --
     .../org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java  | 4 +++-
     1 file changed, 3 insertions(+), 1 deletion(-)
    --
    
    
    http://git-wip-us.apache.org/repos/asf/hbase/blob/4dee4a85/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
    --
    diff --git 
    a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
     
    b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
    index e116f58..9520f5f 100644
    --- 
    a/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
    +++ 
    b/hbase-rsgroup/src/main/java/org/apache/hadoop/hbase/rsgroup/RSGroupInfoManagerImpl.java
    @@ -691,7 +691,9 @@ class RSGroupInfoManagerImpl implements RSGroupInfoManager {
     if(cell != null) {
       sn = 
    ServerName.parseVersionedServerName(CellUtil.cloneValue(cell));
     }
    -if (tsm.isTableState(TableName.NAMESPACE_TABLE_NAME,
    +if (sn == null) {
    +  nsFound.set(false);
    +} else if (tsm.isTableState(TableName.NAMESPACE_TABLE_NAME,
     TableState.State.ENABLED)) {
       try {
     ClientProtos.ClientService.BlockingInterface rs = 
    conn.getClient(sn);