[03/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/testdevapidocs/src-html/org/apache/hadoop/hbase/filter/TestFuzzyRowFilterEndToEnd.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/filter/TestFuzzyRowFilterEndToEnd.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/filter/TestFuzzyRowFilterEndToEnd.html index baa975f..36e14c0 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/filter/TestFuzzyRowFilterEndToEnd.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/filter/TestFuzzyRowFilterEndToEnd.html @@ -25,45 +25,45 @@ 017 */ 018package org.apache.hadoop.hbase.filter; 019 -020import static org.junit.Assert.assertEquals; -021 -022import java.io.IOException; -023import java.nio.ByteBuffer; -024import java.util.ArrayList; -025import java.util.List; -026 -027import org.apache.commons.logging.Log; -028import org.apache.commons.logging.LogFactory; -029import org.apache.hadoop.conf.Configuration; -030import org.apache.hadoop.hbase.Cell; -031import org.apache.hadoop.hbase.CellUtil; -032import org.apache.hadoop.hbase.HBaseTestingUtility; -033import org.apache.hadoop.hbase.HConstants; -034import org.apache.hadoop.hbase.TableName; -035import org.apache.hadoop.hbase.client.Durability; -036import org.apache.hadoop.hbase.client.Put; -037import org.apache.hadoop.hbase.client.Result; -038import org.apache.hadoop.hbase.client.ResultScanner; -039import org.apache.hadoop.hbase.client.Scan; -040import org.apache.hadoop.hbase.client.Table; -041import org.apache.hadoop.hbase.filter.FilterList.Operator; -042import org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy; -043import org.apache.hadoop.hbase.regionserver.HRegion; -044import org.apache.hadoop.hbase.regionserver.RegionScanner; -045import org.apache.hadoop.hbase.testclassification.FilterTests; -046import org.apache.hadoop.hbase.testclassification.MediumTests; -047import org.apache.hadoop.hbase.util.Bytes; -048import org.apache.hadoop.hbase.util.Pair; -049import org.junit.After; -050import org.junit.AfterClass; -051import org.junit.Before; -052import org.junit.BeforeClass; -053import org.junit.Rule; -054import org.junit.Test; -055import org.junit.experimental.categories.Category; -056 -057import org.apache.hadoop.hbase.shaded.com.google.common.collect.Lists; -058import org.junit.rules.TestName; +020import org.apache.commons.logging.Log; +021import org.apache.commons.logging.LogFactory; +022import org.apache.hadoop.conf.Configuration; +023import org.apache.hadoop.hbase.Cell; +024import org.apache.hadoop.hbase.CellUtil; +025import org.apache.hadoop.hbase.HBaseTestingUtility; +026import org.apache.hadoop.hbase.HConstants; +027import org.apache.hadoop.hbase.TableName; +028import org.apache.hadoop.hbase.client.Durability; +029import org.apache.hadoop.hbase.client.Put; +030import org.apache.hadoop.hbase.client.Result; +031import org.apache.hadoop.hbase.client.ResultScanner; +032import org.apache.hadoop.hbase.client.Scan; +033import org.apache.hadoop.hbase.client.Table; +034import org.apache.hadoop.hbase.filter.FilterList.Operator; +035import org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy; +036import org.apache.hadoop.hbase.regionserver.HRegion; +037import org.apache.hadoop.hbase.regionserver.RegionScanner; +038import org.apache.hadoop.hbase.shaded.com.google.common.collect.Lists; +039import org.apache.hadoop.hbase.testclassification.FilterTests; +040import org.apache.hadoop.hbase.testclassification.MediumTests; +041import org.apache.hadoop.hbase.util.Bytes; +042import org.apache.hadoop.hbase.util.Pair; +043import org.junit.After; +044import org.junit.AfterClass; +045import org.junit.Before; +046import org.junit.BeforeClass; +047import org.junit.Rule; +048import org.junit.Test; +049import org.junit.experimental.categories.Category; +050import org.junit.rules.TestName; +051 +052import java.io.IOException; +053import java.nio.ByteBuffer; +054import java.util.ArrayList; +055import java.util.Arrays; +056import java.util.List; +057 +058import static org.junit.Assert.assertEquals; 059 060/** 061 */ @@ -148,311 +148,320 @@ 140ListPairbyte[], byte[] data = new ArrayList(); 141byte[] fuzzyKey = Bytes.toBytesBinary("\\x9B\\x00\\x044e"); 142byte[] mask = new byte[] { 0, 0, 0, 0, 0 }; -143data.add(new Pair(fuzzyKey, mask)); -144FuzzyRowFilter filter = new FuzzyRowFilter(data); -145 -146Scan scan = new Scan(); -147scan.setFilter(filter); -148 -149ResultScanner scanner = ht.getScanner(scan); -150int total = 0; -151while (scanner.next() != null) { -152 total++; -153} -154assertEquals(2, total); -155 TEST_UTIL.deleteTable(TableName.valueOf(name.getMethodName())); -156 } -157 -158 @Test -159 public void testHBASE14782() throws IOException -160 { -161String cf = "f"; -162String cq = "q"; +143 +144// copy
[17/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.SatisfiesCode.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.SatisfiesCode.html b/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.SatisfiesCode.html index a310286..e6ec67b 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.SatisfiesCode.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.SatisfiesCode.html @@ -84,552 +84,560 @@ 076 private RowTracker tracker; 077 078 public FuzzyRowFilter(ListPairbyte[], byte[] fuzzyKeysData) { -079Pairbyte[], byte[] p; -080for (int i = 0; i fuzzyKeysData.size(); i++) { -081 p = fuzzyKeysData.get(i); -082 if (p.getFirst().length != p.getSecond().length) { +079ListPairbyte[], byte[] fuzzyKeyDataCopy = new ArrayList(fuzzyKeysData.size()); +080 +081for (Pairbyte[], byte[] aFuzzyKeysData : fuzzyKeysData) { +082 if (aFuzzyKeysData.getFirst().length != aFuzzyKeysData.getSecond().length) { 083PairString, String readable = -084new Pair(Bytes.toStringBinary(p.getFirst()), Bytes.toStringBinary(p -085.getSecond())); -086throw new IllegalArgumentException("Fuzzy pair lengths do not match: " + readable); -087 } -088 // update mask ( 0 - -1 (0xff), 1 - 2) -089 p.setSecond(preprocessMask(p.getSecond())); -090 preprocessSearchKey(p); -091} -092this.fuzzyKeysData = fuzzyKeysData; -093this.tracker = new RowTracker(); -094 } -095 -096 private void preprocessSearchKey(Pairbyte[], byte[] p) { -097if (!UNSAFE_UNALIGNED) { -098 // do nothing -099 return; -100} -101byte[] key = p.getFirst(); -102byte[] mask = p.getSecond(); -103for (int i = 0; i mask.length; i++) { -104 // set non-fixed part of a search key to 0. -105 if (mask[i] == 2) { -106key[i] = 0; -107 } +084 new Pair(Bytes.toStringBinary(aFuzzyKeysData.getFirst()), Bytes.toStringBinary(aFuzzyKeysData.getSecond())); +085throw new IllegalArgumentException("Fuzzy pair lengths do not match: " + readable); +086 } +087 +088 Pairbyte[], byte[] p = new Pair(); +089 // create a copy of pair bytes so that they are not modified by the filter. +090 p.setFirst(Arrays.copyOf(aFuzzyKeysData.getFirst(), aFuzzyKeysData.getFirst().length)); +091 p.setSecond(Arrays.copyOf(aFuzzyKeysData.getSecond(), aFuzzyKeysData.getSecond().length)); +092 +093 // update mask ( 0 - -1 (0xff), 1 - 2) +094 p.setSecond(preprocessMask(p.getSecond())); +095 preprocessSearchKey(p); +096 +097 fuzzyKeyDataCopy.add(p); +098} +099this.fuzzyKeysData = fuzzyKeyDataCopy; +100this.tracker = new RowTracker(); +101 } +102 +103 +104 private void preprocessSearchKey(Pairbyte[], byte[] p) { +105if (!UNSAFE_UNALIGNED) { +106 // do nothing +107 return; 108} -109 } -110 -111 /** -112 * We need to preprocess mask array, as since we treat 2's as unfixed positions and -1 (0xff) as -113 * fixed positions -114 * @param mask -115 * @return mask array -116 */ -117 private byte[] preprocessMask(byte[] mask) { -118if (!UNSAFE_UNALIGNED) { -119 // do nothing -120 return mask; -121} -122if (isPreprocessedMask(mask)) return mask; -123for (int i = 0; i mask.length; i++) { -124 if (mask[i] == 0) { -125mask[i] = -1; // 0 - -1 -126 } else if (mask[i] == 1) { -127mask[i] = 2;// 1 - 2 -128 } +109byte[] key = p.getFirst(); +110byte[] mask = p.getSecond(); +111for (int i = 0; i mask.length; i++) { +112 // set non-fixed part of a search key to 0. +113 if (mask[i] == 2) { +114key[i] = 0; +115 } +116} +117 } +118 +119 /** +120 * We need to preprocess mask array, as since we treat 2's as unfixed positions and -1 (0xff) as +121 * fixed positions +122 * @param mask +123 * @return mask array +124 */ +125 private byte[] preprocessMask(byte[] mask) { +126if (!UNSAFE_UNALIGNED) { +127 // do nothing +128 return mask; 129} -130return mask; -131 } -132 -133 private boolean isPreprocessedMask(byte[] mask) { -134for (int i = 0; i mask.length; i++) { -135 if (mask[i] != -1 mask[i] != 2) { -136return false; -137 } -138} -139return true; -140 } -141 -142 @Override -143 public ReturnCode filterKeyValue(Cell c) { -144final int startIndex = lastFoundIndex = 0 ? lastFoundIndex : 0; -145final int size = fuzzyKeysData.size(); -146for (int i = startIndex; i size + startIndex; i++) { -147 final int index = i % size; -148 Pairbyte[], byte[] fuzzyData = fuzzyKeysData.get(index); -149 // This shift is idempotent - always
[26/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/book.html -- diff --git a/book.html b/book.html index d004c26..570c87b 100644 --- a/book.html +++ b/book.html @@ -30363,7 +30363,69 @@ However any substantive discussion (as with any off-list project-related discuss -152.9. Development of HBase-related Maven archetypes +152.9. The hbase-thirdparty dependency and shading/relocation + +A new project was created for the release of hbase-2.0.0. It was called +hbase-thirdparty. This project exists only to provide the main hbase +project with relocatedor shadedversions of popular thirdparty +libraries such as guava, netty, and protobuf. The mainline HBase project +relies on the relocated versions of these libraries gotten from hbase-thirdparty +rather than on finding these classes in their usual locations. We do this so +we can specify whatever the version we wish. If we dont relocate, we must +harmonize our version to match that which hadoop and/or spark uses. + + +For developers, this means you need to be careful referring to classes from +netty, guava, protobuf, gson, etc. (see the hbase-thirdparty pom.xml for what +it provides). Devs must refer to the hbase-thirdparty provided classes. In +practice, this is usually not an issue (though it can be a bit of a pain). You +will have to hunt for the relocated version of your particular class. Youll +find it by prepending the general relocation prefix of org.apache.hadoop.hbase.shaded.. +For example if you are looking for com.google.protobuf.Message, the relocated +version used by HBase internals can be found at +org.apache.hadoop.hbase.shaded.com.google.protobuf.Message. + + +For a few thirdparty libs, like protobuf (see the protobuf chapter in this book +for the why), your IDE may give you both optionsthe com.google.protobuf. +and the org.apache.hadoop.hbase.shaded.com.google.protobuf.because both +classes are on your CLASSPATH. Unless you are doing the particular juggling +required in Coprocessor Endpoint development (again see above cited protobuf +chapter), youll want to use the shaded version, always. + + +Of note, the relocation of netty is particular. The netty folks have put in +place facility to aid relocation; it seems like shading netty is a popular project. +One case of this requires the setting of a peculiar system property on the JVM +so that classes out in the bundld shared library (.so) can be found in their +relocated location. Here is the property that needs to be set: + + +-Dorg.apache.hadoop.hbase.shaded.io.netty.packagePrefix=org.apache.hadoop.hbase.shaded. + + +(Note that the trailing '.' is required). Starting hbase normally or when running +test suites, the setting of this property is done for you. If you are doing something +out of the ordinary, starting hbase from your own context, youll need to provide +this property on platforms that favor the bundled .so. See release notes on HBASE-18271 +for more. The complaint you see is something like the following: +Cause: java.lang.RuntimeException: Failed construction of Master: class org.apache.hadoop.hbase.master.HMasterorg.apache.hadoop.hbase.shaded.io.netty.channel.epoll. + + +The hbase-thirdparty project has groupid of org.apache.hbase.thirdparty. +As of this writing, it provides three jars; one for netty with an artifactid of +hbase-thirdparty-netty, one for protobuf at hbase-thirdparty-protobuf and then +a jar for all elsegson, guavaat hbase-thirdpaty-miscellaneous. + + +The hbase-thirdparty artifacts are a product produced by the Apache HBase +project under the aegis of the HBase Project Management Committee. Releases +are done via the usual voting project on the hbase dev mailing list. If issue +in the hbase-thirdparty, use the hbase JIRA and mailing lists to post notice. + + + +152.10. Development of HBase-related Maven archetypes The development of HBase-related Maven archetypes was begun with https://issues.apache.org/jira/browse/HBASE-14876;>HBASE-14876. @@ -35007,7 +35069,7 @@ The server will return cellblocks compressed using this same compressor as long Version 3.0.0-SNAPSHOT -Last updated 2017-08-18 14:29:46 UTC +Last updated 2017-08-19 19:25:51 UTC http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/bulk-loads.html -- diff --git a/bulk-loads.html b/bulk-loads.html index f02a0f4..2330869 100644 --- a/bulk-loads.html +++ b/bulk-loads.html @@ -7,7 +7,7 @@ - + Apache HBase Bulk Loads in Apache HBase (TM) @@ -311,7 +311,7 @@ under the License. --> https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-08-18 + Last Published: 2017-08-19
[28/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/apidocs/index-all.html -- diff --git a/apidocs/index-all.html b/apidocs/index-all.html index 6d0c7b0..2578bcb 100644 --- a/apidocs/index-all.html +++ b/apidocs/index-all.html @@ -2528,6 +2528,8 @@ Create a new namespace +createNextOnRowCol(Cell) - Static method in class org.apache.hadoop.hbase.CellUtil + createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableInputFormatBase Builds a TableRecordReader. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/apidocs/org/apache/hadoop/hbase/class-use/Cell.html -- diff --git a/apidocs/org/apache/hadoop/hbase/class-use/Cell.html b/apidocs/org/apache/hadoop/hbase/class-use/Cell.html index ab72b31..b91d5a5 100644 --- a/apidocs/org/apache/hadoop/hbase/class-use/Cell.html +++ b/apidocs/org/apache/hadoop/hbase/class-use/Cell.html @@ -316,6 +316,10 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. Create a Cell that is larger than all other possible Cells for the given Cell's rk:cf:q. + +static Cell +CellUtil.createNextOnRowCol(Cellcell) + @@ -518,65 +522,69 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. +static Cell +CellUtil.createNextOnRowCol(Cellcell) + + static boolean CellUtil.equals(Cella, Cellb) equals - + static boolean CellUtil.equalsIgnoreMvccVersion(Cella, Cellb) special case for Cell.equals - + static long CellUtil.estimatedHeapSizeOf(Cellcell) This is an estimate of the heap space occupied by a cell. - + static int CellUtil.estimatedSerializedSizeOf(Cellcell) Estimate based on keyvalue's serialization format. - + static int CellUtil.estimatedSerializedSizeOfKey(Cellcell) Calculates the serialized key size. - + static ByteRange CellUtil.fillFamilyRange(Cellcell, ByteRangerange) - + static ByteRange CellUtil.fillQualifierRange(Cellcell, ByteRangerange) - + static ByteRange CellUtil.fillRowRange(Cellcell, ByteRangerange) ByteRange - + static ByteRange CellUtil.fillTagRange(Cellcell, ByteRangerange) - + static ByteRange CellUtil.fillValueRange(Cellcell, ByteRangerange) - + static int CellUtil.findCommonPrefixInFlatKey(Cellc1, Cellc2, @@ -586,17 +594,17 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. KeyValue. - + static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String CellUtil.getCellKeyAsString(Cellcell) - + static byte[] CellUtil.getCellKeySerializedAsKeyValueKey(Cellcell) This method exists just to encapsulate how we serialize keys. - + static http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true; title="class or interface in java.nio">ByteBuffer CellUtil.getQualifierBufferShallowCopy(Cellcell) Deprecated. @@ -604,104 +612,104 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. - + static byte CellUtil.getQualifierByte(Cellcell, intindex) - + static int CellUtil.getRowAsInt(Cellcell) Converts the rowkey bytes of the given cell into an int value - + static byte CellUtil.getRowByte(Cellcell, intindex) misc - + static org.apache.hadoop.hbase.Tag CellUtil.getTag(Cellcell, bytetype) Retrieve Cell's first tag, matching the passed in type - + static byte[] CellUtil.getTagArray(Cellcell) Returns tag value in a new byte array. - + static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.hbase.Tag CellUtil.getTags(Cellcell) - + static http://docs.oracle.com/javase/8/docs/api/java/math/BigDecimal.html?is-external=true; title="class or interface in java.math">BigDecimal CellUtil.getValueAsBigDecimal(Cellcell) Converts the value bytes of the given cell into a BigDecimal - + static double CellUtil.getValueAsDouble(Cellcell) Converts the value bytes of the given cell into a double value - + static int CellUtil.getValueAsInt(Cellcell) Converts the value bytes of the given cell into a int value - + static long CellUtil.getValueAsLong(Cellcell) Converts the value bytes of the given cell into a long value - + static http://docs.oracle.com/javase/8/docs/api/java/nio/ByteBuffer.html?is-external=true; title="class or interface in java.nio">ByteBuffer CellUtil.getValueBufferShallowCopy(Cellcell) - + static boolean CellUtil.isDelete(Cellcell) - + static boolean CellUtil.isDeleteColumnOrFamily(Cellcell) - + static boolean
[25/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/checkstyle-aggregate.html -- diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html index e627084..4a7c971 100644 --- a/checkstyle-aggregate.html +++ b/checkstyle-aggregate.html @@ -7,7 +7,7 @@ - + Apache HBase Checkstyle Results @@ -289,7 +289,7 @@ 2025 0 0 -12858 +12816 Files @@ -1852,7 +1852,7 @@ org/apache/hadoop/hbase/filter/FuzzyRowFilter.java 0 0 -16 +17 org/apache/hadoop/hbase/filter/InclusiveStopFilter.java 0 @@ -4872,7 +4872,7 @@ org/apache/hadoop/hbase/regionserver/querymatcher/ScanQueryMatcher.java 0 0 -3 +4 org/apache/hadoop/hbase/regionserver/snapshot/RegionServerSnapshotManager.java 0 @@ -5892,7 +5892,7 @@ org/apache/hadoop/hbase/security/visibility/VisibilityScanDeleteTracker.java 0 0 -49 +5 org/apache/hadoop/hbase/security/visibility/VisibilityUtils.java 0 @@ -7104,7 +7104,7 @@ caseIndent: 2 basicOffset: 2 lineWrappingIndentation: 2 -3833 +3787 Error javadoc @@ -7134,7 +7134,7 @@ max: 100 ignorePattern: ^package.*|^import.*|a href|href|http://|https://|ftp://|org.apache.thrift.|com.google.protobuf.|hbase.protobuf.generated -960 +964 Error @@ -27743,158 +27743,164 @@ 40 Error +sizes +LineLength +Line is longer than 100 characters (found 120). +84 + +Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -114 - +122 + Error blocks NeedBraces 'if' construct must use '{}'s. -122 - +130 + Error indentation Indentation 'method def modifier' have incorrect indentation level 16, expected level should be one of the following: 10, 12. -194 - +202 + Error indentation Indentation 'method def' child have incorrect indentation level 18, expected level should be one of the following: 12, 14. -197 - +205 + Error indentation Indentation 'method def rcurly' have incorrect indentation level 16, expected level should be one of the following: 10, 12. -199 - +207 + Error indentation Indentation 'object def rcurly' have incorrect indentation level 14, expected level should be one of the following: 8, 10. -200 - +208 + Error sizes LineLength Line is longer than 100 characters (found 107). -206 - -Error -javadoc -NonEmptyAtclauseDescription -At-clause should have a non-empty description. -267 +214 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -586 +275 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -587 +594 Error -blocks -NeedBraces -'if' construct must use '{}'s. -596 +javadoc +NonEmptyAtclauseDescription +At-clause should have a non-empty description. +595 Error blocks NeedBraces 'if' construct must use '{}'s. -598 +604 Error blocks NeedBraces 'if' construct must use '{}'s. -609 +606 Error blocks NeedBraces 'if' construct must use '{}'s. -610 +617 Error blocks NeedBraces 'if' construct must use '{}'s. -613 +618 + +Error +blocks +NeedBraces +'if' construct must use '{}'s. +621 org/apache/hadoop/hbase/filter/InclusiveStopFilter.java - + Severity Category Rule Message Line - + Error imports ImportOrder Wrong order for 'org.apache.hadoop.hbase.shaded.com.google.common.base.Preconditions' import. 33 - + Error blocks NeedBraces 'if' construct must use '{}'s. 56 - + Error blocks NeedBraces 'if' construct must use '{}'s. 62 - + Error whitespace MethodParamPad '(' is preceded with whitespace. 72 - + Error blocks NeedBraces 'if' construct must use '{}'s. 85 - + Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. 93 - + Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. 108 - + Error javadoc JavadocTagContinuationIndentation Line continuation have incorrect indentation level, expected level should be 2. 110 - + Error blocks NeedBraces 'if' construct must use '{}'s. 113 - + Error blocks NeedBraces @@ -27903,55 +27909,55 @@ org/apache/hadoop/hbase/filter/KeyOnlyFilter.java - + Severity Category Rule Message Line - + Error imports ImportOrder Wrong order for 'org.apache.hadoop.hbase.shaded.com.google.common.base.Preconditions' import. 34 - + Error blocks LeftCurly '{' at column 26 should have line break after. 48 - + Error blocks LeftCurly '{' at column 42 should have line break after. 49 - + Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. 98 - + Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. 113 - + Error javadoc JavadocTagContinuationIndentation Line continuation have incorrect indentation level, expected level should be 2. 115 - + Error blocks NeedBraces 'if' construct must use '{}'s. 118 - + Error blocks NeedBraces @@ -27960,145
[09/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/testdevapidocs/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.DeleteMark.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.DeleteMark.html b/testdevapidocs/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.DeleteMark.html new file mode 100644 index 000..d479d7c --- /dev/null +++ b/testdevapidocs/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.DeleteMark.html @@ -0,0 +1,382 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +TestVisibilityLabelsWithDeletes.DeleteMark (Apache HBase 3.0.0-SNAPSHOT Test API) + + + + + +var methods = {"i0":9,"i1":9}; +var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],8:["t4","Concrete Methods"]}; +var altColor = "altColor"; +var rowColor = "rowColor"; +var tableTab = "tableTab"; +var activeTableTab = "activeTableTab"; + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +PrevClass +NextClass + + +Frames +NoFrames + + +AllClasses + + + + + + + +Summary: +Nested| +Enum Constants| +Field| +Method + + +Detail: +Enum Constants| +Field| +Method + + + + + + + + +org.apache.hadoop.hbase.security.visibility +Enum TestVisibilityLabelsWithDeletes.DeleteMark + + + +http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">java.lang.Object + + +http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">java.lang.EnumTestVisibilityLabelsWithDeletes.DeleteMark + + +org.apache.hadoop.hbase.security.visibility.TestVisibilityLabelsWithDeletes.DeleteMark + + + + + + + + + +All Implemented Interfaces: +http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable, http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableTestVisibilityLabelsWithDeletes.DeleteMark + + +Enclosing class: +TestVisibilityLabelsWithDeletes + + + +private static enum TestVisibilityLabelsWithDeletes.DeleteMark +extends http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumTestVisibilityLabelsWithDeletes.DeleteMark + + + + + + + + + + + +Enum Constant Summary + +Enum Constants + +Enum Constant and Description + + +CELL + + +COLUMN + + +FAMILY + + +FAMILY_VERSION + + +ROW + + + + + + + + + +Method Summary + +All MethodsStatic MethodsConcrete Methods + +Modifier and Type +Method and Description + + +static TestVisibilityLabelsWithDeletes.DeleteMark +valueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) +Returns the enum constant of this type with the specified name. + + + +static TestVisibilityLabelsWithDeletes.DeleteMark[] +values() +Returns an array containing the constants of this enum type, in +the order they are declared. + + + + + + + +Methods inherited from classjava.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">Enum +http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true#clone--; title="class or interface in java.lang">clone, http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true#compareTo-E-; title="class or interface in java.lang">compareTo, http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true#equals-java.lang.Object-; title="class or interface in java.lang">equals, http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true#finalize--; title="class or interface in java.lang">finalize, http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true#getDeclaringClass--; title="class or interface in java.lang">getDeclaringClass, http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true#hashCode--; title="class or interface in java.lang">hashCode, http://docs.oracle.com/javase/8/docs/ api/java/lang/Enum.html?is-external=true#name--" title="class or interface in java.lang">name,
[10/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/testdevapidocs/org/apache/hadoop/hbase/client/TestFromClientSide3.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/client/TestFromClientSide3.html b/testdevapidocs/org/apache/hadoop/hbase/client/TestFromClientSide3.html index 189d5a3..8e639a8 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/client/TestFromClientSide3.html +++ b/testdevapidocs/org/apache/hadoop/hbase/client/TestFromClientSide3.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":9,"i1":9,"i2":9,"i3":10,"i4":10,"i5":10,"i6":9,"i7":10,"i8":9,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10}; +var methods = {"i0":9,"i1":9,"i2":9,"i3":10,"i4":10,"i5":10,"i6":9,"i7":10,"i8":9,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":9}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab"; -public class TestFromClientSide3 +public class TestFromClientSide3 extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object @@ -363,6 +363,18 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? void testRowMutationsWithPreBatchMutate() + +void +testScanAfterDeletingSpecifiedRow() + + +void +testScanAfterDeletingSpecifiedRowV2() + + +private static http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">Listorg.apache.hadoop.hbase.Cell +toList(org.apache.hadoop.hbase.client.ResultScannerscanner) + @@ -391,7 +403,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? LOG -private static finalorg.apache.commons.logging.Log LOG +private static finalorg.apache.commons.logging.Log LOG @@ -400,7 +412,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? TEST_UTIL -private static finalHBaseTestingUtility TEST_UTIL +private static finalHBaseTestingUtility TEST_UTIL @@ -409,7 +421,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? FAMILY -private staticbyte[] FAMILY +private staticbyte[] FAMILY @@ -418,7 +430,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? random -private statichttp://docs.oracle.com/javase/8/docs/api/java/util/Random.html?is-external=true; title="class or interface in java.util">Random random +private statichttp://docs.oracle.com/javase/8/docs/api/java/util/Random.html?is-external=true; title="class or interface in java.util">Random random @@ -427,7 +439,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? SLAVES -private staticint SLAVES +private staticint SLAVES @@ -436,7 +448,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? ROW -private staticbyte[] ROW +private staticbyte[] ROW @@ -445,7 +457,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? ANOTHERROW -private static finalbyte[] ANOTHERROW +private static finalbyte[] ANOTHERROW @@ -454,7 +466,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? QUALIFIER -private staticbyte[] QUALIFIER +private staticbyte[] QUALIFIER @@ -463,7 +475,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? VALUE -private staticbyte[] VALUE +private staticbyte[] VALUE @@ -472,7 +484,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? COL_QUAL -private static finalbyte[] COL_QUAL +private static finalbyte[] COL_QUAL @@ -481,7 +493,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? VAL_BYTES -private static finalbyte[] VAL_BYTES +private static finalbyte[] VAL_BYTES @@ -490,7 +502,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? ROW_BYTES -private static finalbyte[] ROW_BYTES +private static finalbyte[] ROW_BYTES @@ -499,7 +511,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? name -publicorg.junit.rules.TestName name +publicorg.junit.rules.TestName name @@ -516,7 +528,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? TestFromClientSide3 -publicTestFromClientSide3() +publicTestFromClientSide3() @@ -533,7 +545,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? setUpBeforeClass -public staticvoidsetUpBeforeClass()
[14/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/devapidocs/src-html/org/apache/hadoop/hbase/security/visibility/VisibilityController.DeleteVersionVisibilityExpressionFilter.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/security/visibility/VisibilityController.DeleteVersionVisibilityExpressionFilter.html b/devapidocs/src-html/org/apache/hadoop/hbase/security/visibility/VisibilityController.DeleteVersionVisibilityExpressionFilter.html index a8f5704..7230e95 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/security/visibility/VisibilityController.DeleteVersionVisibilityExpressionFilter.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/security/visibility/VisibilityController.DeleteVersionVisibilityExpressionFilter.html @@ -1082,56 +1082,60 @@ 1074public ReturnCode filterKeyValue(Cell cell) throws IOException { 1075 ListTag putVisTags = new ArrayList(); 1076 Byte putCellVisTagsFormat = VisibilityUtils.extractVisibilityTags(cell, putVisTags); -1077 boolean matchFound = VisibilityLabelServiceManager -1078 .getInstance().getVisibilityLabelService() -1079 .matchVisibility(putVisTags, putCellVisTagsFormat, deleteCellVisTags, -1080 deleteCellVisTagsFormat); -1081 return matchFound ? ReturnCode.INCLUDE : ReturnCode.SKIP; -1082} -1083 } -1084 -1085 /** -1086 * A RegionServerObserver impl that provides the custom -1087 * VisibilityReplicationEndpoint. This class should be configured as the -1088 * 'hbase.coprocessor.regionserver.classes' for the visibility tags to be -1089 * replicated as string. The value for the configuration should be -1090 * 'org.apache.hadoop.hbase.security.visibility.VisibilityController$VisibilityReplication'. -1091 */ -1092 public static class VisibilityReplication implements RegionServerObserver { -1093private Configuration conf; -1094private VisibilityLabelService visibilityLabelService; -1095 -1096@Override -1097public void start(CoprocessorEnvironment env) throws IOException { -1098 this.conf = env.getConfiguration(); -1099 visibilityLabelService = VisibilityLabelServiceManager.getInstance() -1100 .getVisibilityLabelService(this.conf); -1101} -1102 -1103@Override -1104public void stop(CoprocessorEnvironment env) throws IOException { +1077 if (putVisTags.isEmpty() deleteCellVisTags.isEmpty()) { +1078// Early out if there are no tags in the cell +1079return ReturnCode.INCLUDE; +1080 } +1081 boolean matchFound = VisibilityLabelServiceManager +1082 .getInstance().getVisibilityLabelService() +1083 .matchVisibility(putVisTags, putCellVisTagsFormat, deleteCellVisTags, +1084 deleteCellVisTagsFormat); +1085 return matchFound ? ReturnCode.INCLUDE : ReturnCode.SKIP; +1086} +1087 } +1088 +1089 /** +1090 * A RegionServerObserver impl that provides the custom +1091 * VisibilityReplicationEndpoint. This class should be configured as the +1092 * 'hbase.coprocessor.regionserver.classes' for the visibility tags to be +1093 * replicated as string. The value for the configuration should be +1094 * 'org.apache.hadoop.hbase.security.visibility.VisibilityController$VisibilityReplication'. +1095 */ +1096 public static class VisibilityReplication implements RegionServerObserver { +1097private Configuration conf; +1098private VisibilityLabelService visibilityLabelService; +1099 +1100@Override +1101public void start(CoprocessorEnvironment env) throws IOException { +1102 this.conf = env.getConfiguration(); +1103 visibilityLabelService = VisibilityLabelServiceManager.getInstance() +1104 .getVisibilityLabelService(this.conf); 1105} 1106 1107@Override -1108public ReplicationEndpoint postCreateReplicationEndPoint( -1109 ObserverContextRegionServerCoprocessorEnvironment ctx, ReplicationEndpoint endpoint) { -1110 return new VisibilityReplicationEndpoint(endpoint, visibilityLabelService); -} -1112 } -1113 -1114 /** -1115 * @param t -1116 * @return NameValuePair of the exception name to stringified version os exception. -1117 */ -1118 // Copied from ResponseConverter and made private. Only used in here. -1119 private static NameBytesPair buildException(final Throwable t) { -1120NameBytesPair.Builder parameterBuilder = NameBytesPair.newBuilder(); -1121 parameterBuilder.setName(t.getClass().getName()); -1122parameterBuilder.setValue( -1123 ByteString.copyFromUtf8(StringUtils.stringifyException(t))); -1124return parameterBuilder.build(); -1125 } -1126} +1108public void stop(CoprocessorEnvironment env) throws IOException { +1109} +1110 +@Override +1112public ReplicationEndpoint postCreateReplicationEndPoint( +1113
[24/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/checkstyle.rss -- diff --git a/checkstyle.rss b/checkstyle.rss index 0b9e922..feb4d18 100644 --- a/checkstyle.rss +++ b/checkstyle.rss @@ -26,7 +26,7 @@ under the License. 2007 - 2017 The Apache Software Foundation File: 2025, - Errors: 12858, + Errors: 12816, Warnings: 0, Infos: 0 @@ -9015,7 +9015,7 @@ under the License. 0 - 3 + 4 @@ -9589,7 +9589,7 @@ under the License. 0 - 16 + 17 @@ -17079,7 +17079,7 @@ under the License. 0 - 49 + 5 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/coc.html -- diff --git a/coc.html b/coc.html index e5aa2d9..493e9bb 100644 --- a/coc.html +++ b/coc.html @@ -7,7 +7,7 @@ - + Apache HBase Code of Conduct Policy @@ -380,7 +380,7 @@ email to mailto:priv...@hbase.apache.org;>the priv https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-08-18 + Last Published: 2017-08-19 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/cygwin.html -- diff --git a/cygwin.html b/cygwin.html index 5250311..50e2494 100644 --- a/cygwin.html +++ b/cygwin.html @@ -7,7 +7,7 @@ - + Apache HBase Installing Apache HBase (TM) on Windows using Cygwin @@ -679,7 +679,7 @@ Now your HBase server is running, start coding and build that next https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-08-18 + Last Published: 2017-08-19 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/dependencies.html -- diff --git a/dependencies.html b/dependencies.html index 5782ec0..c245f96 100644 --- a/dependencies.html +++ b/dependencies.html @@ -7,7 +7,7 @@ - + Apache HBase Project Dependencies @@ -527,7 +527,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-08-18 + Last Published: 2017-08-19 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/dependency-convergence.html -- diff --git a/dependency-convergence.html b/dependency-convergence.html index 4f279dd..876da05 100644 --- a/dependency-convergence.html +++ b/dependency-convergence.html @@ -7,7 +7,7 @@ - + Apache HBase Reactor Dependency Convergence @@ -741,7 +741,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-08-18 + Last Published: 2017-08-19 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/dependency-info.html -- diff --git a/dependency-info.html b/dependency-info.html index 30297e4..4a10b97 100644 --- a/dependency-info.html +++ b/dependency-info.html @@ -7,7 +7,7 @@ - + Apache HBase Dependency Information @@ -318,7 +318,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-08-18 + Last Published: 2017-08-19 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/dependency-management.html -- diff --git a/dependency-management.html b/dependency-management.html index 65862c7..9abf1d7 100644 --- a/dependency-management.html +++ b/dependency-management.html @@ -7,7 +7,7 @@ - +
[07/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/testdevapidocs/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithSLGStack.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithSLGStack.html b/testdevapidocs/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithSLGStack.html index 3df4a8d..253b24e 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithSLGStack.html +++ b/testdevapidocs/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithSLGStack.html @@ -49,7 +49,7 @@ var activeTableTab = "activeTableTab"; -PrevClass +PrevClass NextClass @@ -459,7 +459,7 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? -PrevClass +PrevClass NextClass http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/testdevapidocs/org/apache/hadoop/hbase/security/visibility/class-use/TestVisibilityLabelsWithDeletes.DeleteMark.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/security/visibility/class-use/TestVisibilityLabelsWithDeletes.DeleteMark.html b/testdevapidocs/org/apache/hadoop/hbase/security/visibility/class-use/TestVisibilityLabelsWithDeletes.DeleteMark.html new file mode 100644 index 000..b040f3b --- /dev/null +++ b/testdevapidocs/org/apache/hadoop/hbase/security/visibility/class-use/TestVisibilityLabelsWithDeletes.DeleteMark.html @@ -0,0 +1,201 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + + + + +Uses of Class org.apache.hadoop.hbase.security.visibility.TestVisibilityLabelsWithDeletes.DeleteMark (Apache HBase 3.0.0-SNAPSHOT Test API) + + + + + + + +JavaScript is disabled on your browser. + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +Prev +Next + + +Frames +NoFrames + + +AllClasses + + + + + + + + + + +Uses of Classorg.apache.hadoop.hbase.security.visibility.TestVisibilityLabelsWithDeletes.DeleteMark + + + + + +Packages that use TestVisibilityLabelsWithDeletes.DeleteMark + +Package +Description + + + +org.apache.hadoop.hbase.security.visibility + + + + + + + + + + +Uses of TestVisibilityLabelsWithDeletes.DeleteMark in org.apache.hadoop.hbase.security.visibility + +Methods in org.apache.hadoop.hbase.security.visibility that return TestVisibilityLabelsWithDeletes.DeleteMark + +Modifier and Type +Method and Description + + + +static TestVisibilityLabelsWithDeletes.DeleteMark +TestVisibilityLabelsWithDeletes.DeleteMark.valueOf(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) +Returns the enum constant of this type with the specified name. + + + +static TestVisibilityLabelsWithDeletes.DeleteMark[] +TestVisibilityLabelsWithDeletes.DeleteMark.values() +Returns an array containing the constants of this enum type, in +the order they are declared. + + + + + +Methods in org.apache.hadoop.hbase.security.visibility with parameters of type TestVisibilityLabelsWithDeletes.DeleteMark + +Modifier and Type +Method and Description + + + +private static org.apache.hadoop.hbase.client.Delete +TestVisibilityLabelsWithDeletes.addDeleteMark(org.apache.hadoop.hbase.client.Deleted, + TestVisibilityLabelsWithDeletes.DeleteMarkmark, + longnow) + + +private void +TestVisibilityLabelsWithDeletes.testDeleteCellWithoutVisibility(TestVisibilityLabelsWithDeletes.DeleteMarkmark) + + +private void +TestVisibilityLabelsWithDeletes.testDeleteCellWithVisibility(TestVisibilityLabelsWithDeletes.DeleteMarkmark) + + +private void +TestVisibilityLabelsWithDeletes.testDeleteCellWithVisibilityV2(TestVisibilityLabelsWithDeletes.DeleteMarkmark) + + + + + + + + + + + + +Skip navigation links + + + + +Overview +Package +Class +Use +Tree +Deprecated +Index +Help + + + + +Prev +Next + + +Frames +NoFrames + + +AllClasses + + + + + + + + + +Copyright 20072017 https://www.apache.org/;>The Apache Software Foundation. All rights reserved. + +
[02/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/testdevapidocs/src-html/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.DeleteMark.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.DeleteMark.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.DeleteMark.html new file mode 100644 index 000..f4dc862 --- /dev/null +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.DeleteMark.html @@ -0,0 +1,3551 @@ +http://www.w3.org/TR/html4/loose.dtd;> + + +Source code + + + + +001/** +002 * Licensed to the Apache Software Foundation (ASF) under one +003 * or more contributor license agreements. See the NOTICE file +004 * distributed with this work for additional information +005 * regarding copyright ownership. The ASF licenses this file +006 * to you under the Apache License, Version 2.0 (the +007 * "License"); you may not use this file except in compliance +008 * with the License. You may obtain a copy of the License at +009 * +010 * http://www.apache.org/licenses/LICENSE-2.0 +011 * +012 * Unless required by applicable law or agreed to in writing, software +013 * distributed under the License is distributed on an "AS IS" BASIS, +014 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +015 * See the License for the specific language governing permissions and +016 * limitations under the License. +017 */ +018package org.apache.hadoop.hbase.security.visibility; +019 +020import org.apache.commons.logging.Log; +021import org.apache.commons.logging.LogFactory; +022import org.apache.hadoop.conf.Configuration; +023import org.apache.hadoop.hbase.Cell; +024import org.apache.hadoop.hbase.CellScanner; +025import org.apache.hadoop.hbase.CellUtil; +026import org.apache.hadoop.hbase.HBaseTestingUtility; +027import org.apache.hadoop.hbase.HColumnDescriptor; +028import org.apache.hadoop.hbase.HConstants; +029import org.apache.hadoop.hbase.HTableDescriptor; +030import org.apache.hadoop.hbase.TableName; +031import org.apache.hadoop.hbase.client.Admin; +032import org.apache.hadoop.hbase.client.Connection; +033import org.apache.hadoop.hbase.client.ConnectionFactory; +034import org.apache.hadoop.hbase.client.Delete; +035import org.apache.hadoop.hbase.client.Get; +036import org.apache.hadoop.hbase.client.Put; +037import org.apache.hadoop.hbase.client.Result; +038import org.apache.hadoop.hbase.client.ResultScanner; +039import org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException; +040import org.apache.hadoop.hbase.client.Scan; +041import org.apache.hadoop.hbase.client.Table; +042import org.apache.hadoop.hbase.protobuf.generated.VisibilityLabelsProtos.VisibilityLabelsResponse; +043import org.apache.hadoop.hbase.security.User; +044import org.apache.hadoop.hbase.testclassification.MediumTests; +045import org.apache.hadoop.hbase.testclassification.SecurityTests; +046import org.apache.hadoop.hbase.util.Bytes; +047import org.apache.hadoop.hbase.util.DefaultEnvironmentEdge; +048import org.apache.hadoop.hbase.util.EnvironmentEdge; +049import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; +050import org.apache.hadoop.hbase.util.Threads; +051import org.junit.After; +052import org.junit.AfterClass; +053import org.junit.BeforeClass; +054import org.junit.Ignore; +055import org.junit.Rule; +056import org.junit.Test; +057import org.junit.experimental.categories.Category; +058import org.junit.rules.TestName; +059 +060import java.io.IOException; +061import java.io.InterruptedIOException; +062import java.security.PrivilegedExceptionAction; +063import java.util.ArrayList; +064import java.util.Arrays; +065import java.util.List; +066 +067import static org.apache.hadoop.hbase.security.visibility.VisibilityConstants.LABELS_TABLE_NAME; +068import static org.junit.Assert.assertEquals; +069import static org.junit.Assert.assertTrue; +070 +071/** +072 * Tests visibility labels with deletes +073 */ +074@Category({SecurityTests.class, MediumTests.class}) +075public class TestVisibilityLabelsWithDeletes { +076 private static final Log LOG = LogFactory.getLog(TestVisibilityLabelsWithDeletes.class); +077 private static final String TOPSECRET = "TOPSECRET"; +078 private static final String PUBLIC = "PUBLIC"; +079 private static final String PRIVATE = "PRIVATE"; +080 private static final String CONFIDENTIAL = "CONFIDENTIAL"; +081 private static final String SECRET = "SECRET"; +082 public static final HBaseTestingUtility TEST_UTIL = new HBaseTestingUtility(); +083 private static final byte[] row1 = Bytes.toBytes("row1"); +084 private static final byte[] row2 = Bytes.toBytes("row2"); +085 protected final static byte[] fam = Bytes.toBytes("info"); +086 protected
[20/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.TagRewriteCell.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.TagRewriteCell.html b/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.TagRewriteCell.html index 18d00df..3e734a6 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.TagRewriteCell.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.TagRewriteCell.html @@ -3143,7 +3143,60 @@ 3135 return Type.DeleteFamily.getCode(); 3136} 3137 } -3138} +3138 +3139 /** +3140 * @return An new cell is located following input cell. If both of type and timestamp are +3141 * minimum, the input cell will be returned directly. +3142 */ +3143 @InterfaceAudience.Private +3144 public static Cell createNextOnRowCol(Cell cell) { +3145long ts = cell.getTimestamp(); +3146byte type = cell.getTypeByte(); +3147if (type != Type.Minimum.getCode()) { +3148 type = KeyValue.Type.values()[KeyValue.Type.codeToType(type).ordinal() - 1].getCode(); +3149} else if (ts != HConstants.OLDEST_TIMESTAMP) { +3150 ts = ts - 1; +3151 type = Type.Maximum.getCode(); +3152} else { +3153 return cell; +3154} +3155return createNextOnRowCol(cell, ts, type); +3156 } +3157 +3158 private static Cell createNextOnRowCol(Cell cell, long ts, byte type) { +3159if (cell instanceof ByteBufferCell) { +3160 return new LastOnRowColByteBufferCell(((ByteBufferCell) cell).getRowByteBuffer(), +3161 ((ByteBufferCell) cell).getRowPosition(), cell.getRowLength(), +3162 ((ByteBufferCell) cell).getFamilyByteBuffer(), +3163 ((ByteBufferCell) cell).getFamilyPosition(), cell.getFamilyLength(), +3164 ((ByteBufferCell) cell).getQualifierByteBuffer(), +3165 ((ByteBufferCell) cell).getQualifierPosition(), cell.getQualifierLength()) { +3166@Override +3167public long getTimestamp() { +3168 return ts; +3169} +3170 +3171@Override +3172public byte getTypeByte() { +3173 return type; +3174} +3175 }; +3176} +3177return new LastOnRowColCell(cell.getRowArray(), cell.getRowOffset(), cell.getRowLength(), +3178cell.getFamilyArray(), cell.getFamilyOffset(), cell.getFamilyLength(), +3179cell.getQualifierArray(), cell.getQualifierOffset(), cell.getQualifierLength()) { +3180 @Override +3181 public long getTimestamp() { +3182return ts; +3183 } +3184 +3185 @Override +3186 public byte getTypeByte() { +3187return type; +3188 } +3189}; +3190 } +3191} http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.ValueAndTagRewriteByteBufferCell.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.ValueAndTagRewriteByteBufferCell.html b/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.ValueAndTagRewriteByteBufferCell.html index 18d00df..3e734a6 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.ValueAndTagRewriteByteBufferCell.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.ValueAndTagRewriteByteBufferCell.html @@ -3143,7 +3143,60 @@ 3135 return Type.DeleteFamily.getCode(); 3136} 3137 } -3138} +3138 +3139 /** +3140 * @return An new cell is located following input cell. If both of type and timestamp are +3141 * minimum, the input cell will be returned directly. +3142 */ +3143 @InterfaceAudience.Private +3144 public static Cell createNextOnRowCol(Cell cell) { +3145long ts = cell.getTimestamp(); +3146byte type = cell.getTypeByte(); +3147if (type != Type.Minimum.getCode()) { +3148 type = KeyValue.Type.values()[KeyValue.Type.codeToType(type).ordinal() - 1].getCode(); +3149} else if (ts != HConstants.OLDEST_TIMESTAMP) { +3150 ts = ts - 1; +3151 type = Type.Maximum.getCode(); +3152} else { +3153 return cell; +3154} +3155return createNextOnRowCol(cell, ts, type); +3156 } +3157 +3158 private static Cell createNextOnRowCol(Cell cell, long ts, byte type) { +3159if (cell instanceof ByteBufferCell) { +3160 return new LastOnRowColByteBufferCell(((ByteBufferCell) cell).getRowByteBuffer(), +3161 ((ByteBufferCell) cell).getRowPosition(), cell.getRowLength(), +3162 ((ByteBufferCell) cell).getFamilyByteBuffer(), +3163 ((ByteBufferCell) cell).getFamilyPosition(), cell.getFamilyLength(), +3164 ((ByteBufferCell) cell).getQualifierByteBuffer(), +3165 ((ByteBufferCell) cell).getQualifierPosition(), cell.getQualifierLength()) { +3166@Override +3167
[05/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestFromClientSide3.WatiingForScanObserver.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestFromClientSide3.WatiingForScanObserver.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestFromClientSide3.WatiingForScanObserver.html index 401b413..6bb0bed 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestFromClientSide3.WatiingForScanObserver.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestFromClientSide3.WatiingForScanObserver.html @@ -42,880 +42,981 @@ 034import org.apache.commons.logging.LogFactory; 035import org.apache.hadoop.conf.Configuration; 036import org.apache.hadoop.hbase.Cell; -037import org.apache.hadoop.hbase.Coprocessor; -038import org.apache.hadoop.hbase.coprocessor.ObserverContext; -039import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment; -040import org.apache.hadoop.hbase.coprocessor.RegionObserver; -041import org.apache.hadoop.hbase.regionserver.HRegionServer; -042import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress; -043import org.apache.hadoop.hbase.TableName; -044import org.apache.hadoop.hbase.regionserver.Region; -045import org.apache.hadoop.hbase.regionserver.RegionScanner; -046import org.apache.hadoop.hbase.HBaseTestingUtility; -047import org.apache.hadoop.hbase.HColumnDescriptor; -048import org.apache.hadoop.hbase.HConstants; -049import org.apache.hadoop.hbase.HRegionLocation; -050import org.apache.hadoop.hbase.HTableDescriptor; -051import org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint; -052import org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils; -053import org.apache.hadoop.hbase.ipc.ServerRpcController; -054import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos; -055import org.apache.hadoop.hbase.regionserver.HRegion; -056import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -057import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos; -058import org.apache.hadoop.hbase.testclassification.ClientTests; -059import org.apache.hadoop.hbase.testclassification.LargeTests; -060import org.apache.hadoop.hbase.util.Bytes; -061import org.apache.hadoop.hbase.util.Pair; -062import org.junit.After; -063import org.junit.AfterClass; -064import static org.junit.Assert.assertEquals; -065import static org.junit.Assert.assertFalse; -066import static org.junit.Assert.assertNull; -067import static org.junit.Assert.assertTrue; -068import static org.junit.Assert.fail; -069import org.junit.Before; -070import org.junit.BeforeClass; -071import org.junit.Rule; -072import org.junit.Test; -073import org.junit.experimental.categories.Category; -074import org.junit.rules.TestName; -075 -076@Category({LargeTests.class, ClientTests.class}) -077public class TestFromClientSide3 { -078 private static final Log LOG = LogFactory.getLog(TestFromClientSide3.class); -079 private final static HBaseTestingUtility TEST_UTIL -080= new HBaseTestingUtility(); -081 private static byte[] FAMILY = Bytes.toBytes("testFamily"); -082 private static Random random = new Random(); -083 private static int SLAVES = 3; -084 private static byte [] ROW = Bytes.toBytes("testRow"); -085 private static final byte[] ANOTHERROW = Bytes.toBytes("anotherrow"); -086 private static byte [] QUALIFIER = Bytes.toBytes("testQualifier"); -087 private static byte [] VALUE = Bytes.toBytes("testValue"); -088 private final static byte[] COL_QUAL = Bytes.toBytes("f1"); -089 private final static byte[] VAL_BYTES = Bytes.toBytes("v1"); -090 private final static byte[] ROW_BYTES = Bytes.toBytes("r1"); -091 -092 @Rule -093 public TestName name = new TestName(); -094 -095 /** -096 * @throws java.lang.Exception -097 */ -098 @BeforeClass -099 public static void setUpBeforeClass() throws Exception { -100TEST_UTIL.startMiniCluster(SLAVES); -101 } -102 -103 /** -104 * @throws java.lang.Exception -105 */ -106 @AfterClass -107 public static void tearDownAfterClass() throws Exception { -108TEST_UTIL.shutdownMiniCluster(); -109 } -110 -111 /** -112 * @throws java.lang.Exception -113 */ -114 @Before -115 public void setUp() throws Exception { -116// Nothing to do. -117 } -118 -119 /** -120 * @throws java.lang.Exception -121 */ -122 @After -123 public void tearDown() throws Exception { -124for (HTableDescriptor htd: TEST_UTIL.getAdmin().listTables()) { -125 LOG.info("Tear down, remove table=" + htd.getTableName()); -126 TEST_UTIL.deleteTable(htd.getTableName()); -127 } +037import org.apache.hadoop.hbase.CellUtil; +038import org.apache.hadoop.hbase.Coprocessor; +039import org.apache.hadoop.hbase.coprocessor.ObserverContext; +040import
[22/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/devapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html b/devapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html index 566b30f..e11d72d 100644 --- a/devapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/procedure2/package-tree.html @@ -201,11 +201,11 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.procedure2.LockInfo.ResourceType -org.apache.hadoop.hbase.procedure2.RootProcedureState.State org.apache.hadoop.hbase.procedure2.StateMachineProcedure.Flow -org.apache.hadoop.hbase.procedure2.Procedure.LockState org.apache.hadoop.hbase.procedure2.LockInfo.LockType +org.apache.hadoop.hbase.procedure2.RootProcedureState.State +org.apache.hadoop.hbase.procedure2.LockInfo.ResourceType +org.apache.hadoop.hbase.procedure2.Procedure.LockState http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html b/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html index acc0e97..7b46e5c 100644 --- a/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/quotas/package-tree.html @@ -201,13 +201,13 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.quotas.ThrottleType -org.apache.hadoop.hbase.quotas.OperationQuota.OperationType -org.apache.hadoop.hbase.quotas.ThrottlingException.Type org.apache.hadoop.hbase.quotas.QuotaSnapshotStore.ViolationState -org.apache.hadoop.hbase.quotas.QuotaType -org.apache.hadoop.hbase.quotas.QuotaScope +org.apache.hadoop.hbase.quotas.ThrottlingException.Type +org.apache.hadoop.hbase.quotas.ThrottleType org.apache.hadoop.hbase.quotas.SpaceViolationPolicy +org.apache.hadoop.hbase.quotas.QuotaScope +org.apache.hadoop.hbase.quotas.QuotaType +org.apache.hadoop.hbase.quotas.OperationQuota.OperationType http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html index 95b44fc..c6e17d3 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html @@ -693,20 +693,20 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection -org.apache.hadoop.hbase.regionserver.FlushType -org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope -org.apache.hadoop.hbase.regionserver.MemStoreCompactor.Action -org.apache.hadoop.hbase.regionserver.Region.Operation org.apache.hadoop.hbase.regionserver.RegionOpeningState -org.apache.hadoop.hbase.regionserver.StoreScanner.StoreScannerCompactionRace org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage -org.apache.hadoop.hbase.regionserver.ScannerContext.NextState org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status -org.apache.hadoop.hbase.regionserver.Region.FlushResult.Result +org.apache.hadoop.hbase.regionserver.StoreScanner.StoreScannerCompactionRace org.apache.hadoop.hbase.regionserver.ScanType -org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType
[01/30] hbase-site git commit: Published site at .
Repository: hbase-site Updated Branches: refs/heads/asf-site 2d30087f7 -> 495ddb865 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/testdevapidocs/src-html/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.html index 09e17cb..f4dc862 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.html @@ -25,3275 +25,3466 @@ 017 */ 018package org.apache.hadoop.hbase.security.visibility; 019 -020import org.apache.hadoop.conf.Configuration; -021import org.apache.hadoop.hbase.Cell; -022import org.apache.hadoop.hbase.CellScanner; -023import org.apache.hadoop.hbase.HBaseTestingUtility; -024import org.apache.hadoop.hbase.HColumnDescriptor; -025import org.apache.hadoop.hbase.HConstants; -026import org.apache.hadoop.hbase.HTableDescriptor; -027import org.apache.hadoop.hbase.TableName; -028import org.apache.hadoop.hbase.client.Admin; -029import org.apache.hadoop.hbase.client.Connection; -030import org.apache.hadoop.hbase.client.ConnectionFactory; -031import org.apache.hadoop.hbase.client.Delete; -032import org.apache.hadoop.hbase.client.Get; -033import org.apache.hadoop.hbase.client.Put; -034import org.apache.hadoop.hbase.client.Result; -035import org.apache.hadoop.hbase.client.ResultScanner; -036import org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException; -037import org.apache.hadoop.hbase.client.Scan; -038import org.apache.hadoop.hbase.client.Table; -039import org.apache.hadoop.hbase.protobuf.generated.VisibilityLabelsProtos.VisibilityLabelsResponse; -040import org.apache.hadoop.hbase.security.User; -041import org.apache.hadoop.hbase.testclassification.MediumTests; -042import org.apache.hadoop.hbase.testclassification.SecurityTests; -043import org.apache.hadoop.hbase.util.Bytes; -044import org.apache.hadoop.hbase.util.Threads; -045import org.junit.After; -046import org.junit.AfterClass; -047import org.junit.BeforeClass; -048import org.junit.Ignore; -049import org.junit.Rule; -050import org.junit.Test; -051import org.junit.experimental.categories.Category; -052import org.junit.rules.TestName; -053 -054import java.io.IOException; -055import java.io.InterruptedIOException; -056import java.security.PrivilegedExceptionAction; -057import java.util.ArrayList; -058import java.util.Arrays; -059import java.util.List; -060 -061import static org.apache.hadoop.hbase.security.visibility.VisibilityConstants.LABELS_TABLE_NAME; -062import static org.junit.Assert.assertEquals; -063import static org.junit.Assert.assertTrue; -064 -065/** -066 * Tests visibility labels with deletes -067 */ -068@Category({SecurityTests.class, MediumTests.class}) -069public class TestVisibilityLabelsWithDeletes { -070 private static final String TOPSECRET = "TOPSECRET"; -071 private static final String PUBLIC = "PUBLIC"; -072 private static final String PRIVATE = "PRIVATE"; -073 private static final String CONFIDENTIAL = "CONFIDENTIAL"; -074 private static final String SECRET = "SECRET"; -075 public static final HBaseTestingUtility TEST_UTIL = new HBaseTestingUtility(); -076 private static final byte[] row1 = Bytes.toBytes("row1"); -077 private static final byte[] row2 = Bytes.toBytes("row2"); -078 protected final static byte[] fam = Bytes.toBytes("info"); -079 protected final static byte[] qual = Bytes.toBytes("qual"); -080 private final static byte[] qual1 = Bytes.toBytes("qual1"); -081 private final static byte[] qual2 = Bytes.toBytes("qual2"); -082 protected final static byte[] value = Bytes.toBytes("value"); -083 private final static byte[] value1 = Bytes.toBytes("value1"); -084 public static Configuration conf; -085 -086 @Rule -087 public final TestName TEST_NAME = new TestName(); -088 public static User SUPERUSER; -089 -090 @BeforeClass -091 public static void setupBeforeClass() throws Exception { -092// setup configuration -093conf = TEST_UTIL.getConfiguration(); -094 conf.setBoolean(HConstants.DISTRIBUTED_LOG_REPLAY_KEY, false); -095 VisibilityTestUtil.enableVisiblityLabels(conf); -096 conf.setClass(VisibilityUtils.VISIBILITY_LABEL_GENERATOR_CLASS, SimpleScanLabelGenerator.class, -097ScanLabelGenerator.class); -098conf.set("hbase.superuser", "admin"); -099TEST_UTIL.startMiniCluster(2); -100SUPERUSER = User.createUserForTesting(conf, "admin", new String[] { "supergroup" }); -101 -102// Wait for the labels table to become available -103 TEST_UTIL.waitTableEnabled(LABELS_TABLE_NAME.getName(), 5); -104
[15/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/querymatcher/ScanQueryMatcher.MatchCode.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/querymatcher/ScanQueryMatcher.MatchCode.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/querymatcher/ScanQueryMatcher.MatchCode.html index cf26433..07c9f18 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/querymatcher/ScanQueryMatcher.MatchCode.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/querymatcher/ScanQueryMatcher.MatchCode.html @@ -298,122 +298,135 @@ 290 public abstract boolean moreRowsMayExistAfter(Cell cell); 291 292 public Cell getKeyForNextColumn(Cell cell) { -293ColumnCount nextColumn = columns.getColumnHint(); -294if (nextColumn == null) { -295 return CellUtil.createLastOnRowCol(cell); -296} else { -297 return CellUtil.createFirstOnRowCol(cell, nextColumn.getBuffer(), nextColumn.getOffset(), -298nextColumn.getLength()); -299} -300 } -301 -302 /** -303 * @param nextIndexed the key of the next entry in the block index (if any) -304 * @param currentCell The Cell we're using to calculate the seek key -305 * @return result of the compare between the indexed key and the key portion of the passed cell -306 */ -307 public int compareKeyForNextRow(Cell nextIndexed, Cell currentCell) { -308return rowComparator.compareKeyBasedOnColHint(nextIndexed, currentCell, 0, 0, null, 0, 0, -309 HConstants.OLDEST_TIMESTAMP, Type.Minimum.getCode()); -310 } -311 -312 /** -313 * @param nextIndexed the key of the next entry in the block index (if any) -314 * @param currentCell The Cell we're using to calculate the seek key -315 * @return result of the compare between the indexed key and the key portion of the passed cell -316 */ -317 public int compareKeyForNextColumn(Cell nextIndexed, Cell currentCell) { -318ColumnCount nextColumn = columns.getColumnHint(); -319if (nextColumn == null) { -320 return rowComparator.compareKeyBasedOnColHint(nextIndexed, currentCell, 0, 0, null, 0, 0, -321HConstants.OLDEST_TIMESTAMP, Type.Minimum.getCode()); -322} else { -323 return rowComparator.compareKeyBasedOnColHint(nextIndexed, currentCell, -324currentCell.getFamilyOffset(), currentCell.getFamilyLength(), nextColumn.getBuffer(), -325nextColumn.getOffset(), nextColumn.getLength(), HConstants.LATEST_TIMESTAMP, -326Type.Maximum.getCode()); -327} -328 } -329 -330 /** -331 * @return the Filter -332 */ -333 public abstract Filter getFilter(); -334 -335 /** -336 * Delegate to {@link Filter#getNextCellHint(Cell)}. If no filter, return {@code null}. -337 */ -338 public abstract Cell getNextKeyHint(Cell cell) throws IOException; -339 -340 @Override -341 public void beforeShipped() throws IOException { -342if (this.currentRow != null) { -343 this.currentRow = CellUtil.createFirstOnRow(CellUtil.copyRow(this.currentRow)); -344} -345if (columns != null) { -346 columns.beforeShipped(); -347} -348 } -349 -350 protected static Cell createStartKeyFromRow(byte[] startRow, ScanInfo scanInfo) { -351return CellUtil.createFirstDeleteFamilyCellOnRow(startRow, scanInfo.getFamily()); -352 } -353 -354 protected static PairDeleteTracker, ColumnTracker getTrackers(RegionCoprocessorHost host, -355 NavigableSetbyte[] columns, ScanInfo scanInfo, long oldestUnexpiredTS, Scan userScan) -356 throws IOException { -357int resultMaxVersion = scanInfo.getMaxVersions(); -358int maxVersionToCheck = resultMaxVersion; -359if (userScan != null) { -360 if (userScan.isRaw()) { -361resultMaxVersion = userScan.getMaxVersions(); -362 } else { -363resultMaxVersion = Math.min(userScan.getMaxVersions(), scanInfo.getMaxVersions()); -364 } -365 maxVersionToCheck = userScan.hasFilter() ? scanInfo.getMaxVersions() : resultMaxVersion; -366} -367 -368DeleteTracker deleteTracker; -369if (scanInfo.isNewVersionBehavior() (userScan == null || !userScan.isRaw())) { -370 deleteTracker = new NewVersionBehaviorTracker(columns, scanInfo.getMinVersions(), -371 scanInfo.getMaxVersions(), resultMaxVersion, oldestUnexpiredTS); -372} else { -373 deleteTracker = new ScanDeleteTracker(); -374} -375if (host != null) { -376 deleteTracker = host.postInstantiateDeleteTracker(deleteTracker); -377 if (deleteTracker instanceof VisibilityScanDeleteTracker scanInfo.isNewVersionBehavior()) { -378deleteTracker = new VisibilityNewVersionBehaivorTracker(columns, scanInfo.getMinVersions(), -379scanInfo.getMaxVersions(), resultMaxVersion, oldestUnexpiredTS); -380 } -381} -382 -383ColumnTracker
[12/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/hbase-archetypes/dependencies.html -- diff --git a/hbase-archetypes/dependencies.html b/hbase-archetypes/dependencies.html index b8ab78f..61bede8 100644 --- a/hbase-archetypes/dependencies.html +++ b/hbase-archetypes/dependencies.html @@ -7,7 +7,7 @@ - + Apache HBase - Archetypes Project Dependencies @@ -354,7 +354,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-08-18 + Last Published: 2017-08-19 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/hbase-archetypes/dependency-convergence.html -- diff --git a/hbase-archetypes/dependency-convergence.html b/hbase-archetypes/dependency-convergence.html index 7aa2da9..dfb0693 100644 --- a/hbase-archetypes/dependency-convergence.html +++ b/hbase-archetypes/dependency-convergence.html @@ -7,7 +7,7 @@ - + Apache HBase - Archetypes Reactor Dependency Convergence @@ -571,7 +571,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-08-18 + Last Published: 2017-08-19 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/hbase-archetypes/dependency-info.html -- diff --git a/hbase-archetypes/dependency-info.html b/hbase-archetypes/dependency-info.html index 69e6077..be0fc4d 100644 --- a/hbase-archetypes/dependency-info.html +++ b/hbase-archetypes/dependency-info.html @@ -7,7 +7,7 @@ - + Apache HBase - Archetypes Dependency Information @@ -148,7 +148,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-08-18 + Last Published: 2017-08-19 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/hbase-archetypes/dependency-management.html -- diff --git a/hbase-archetypes/dependency-management.html b/hbase-archetypes/dependency-management.html index 3e52d97..3b2a88d 100644 --- a/hbase-archetypes/dependency-management.html +++ b/hbase-archetypes/dependency-management.html @@ -7,7 +7,7 @@ - + Apache HBase - Archetypes Project Dependency Management @@ -736,7 +736,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-08-18 + Last Published: 2017-08-19 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/hbase-archetypes/hbase-archetype-builder/dependencies.html -- diff --git a/hbase-archetypes/hbase-archetype-builder/dependencies.html b/hbase-archetypes/hbase-archetype-builder/dependencies.html index 6072376..590f36d 100644 --- a/hbase-archetypes/hbase-archetype-builder/dependencies.html +++ b/hbase-archetypes/hbase-archetype-builder/dependencies.html @@ -7,7 +7,7 @@ - + Apache HBase - Archetype builder Project Dependencies @@ -354,7 +354,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-08-18 + Last Published: 2017-08-19 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/hbase-archetypes/hbase-archetype-builder/dependency-convergence.html -- diff --git a/hbase-archetypes/hbase-archetype-builder/dependency-convergence.html b/hbase-archetypes/hbase-archetype-builder/dependency-convergence.html index fe92e06..719bcbe 100644 --- a/hbase-archetypes/hbase-archetype-builder/dependency-convergence.html +++ b/hbase-archetypes/hbase-archetype-builder/dependency-convergence.html @@ -7,7 +7,7 @@ - + Apache HBase - Archetype builder Reactor Dependency Convergence @@ -571,7 +571,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-08-18 +
[16/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html b/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html index a310286..e6ec67b 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html @@ -84,552 +84,560 @@ 076 private RowTracker tracker; 077 078 public FuzzyRowFilter(ListPairbyte[], byte[] fuzzyKeysData) { -079Pairbyte[], byte[] p; -080for (int i = 0; i fuzzyKeysData.size(); i++) { -081 p = fuzzyKeysData.get(i); -082 if (p.getFirst().length != p.getSecond().length) { +079ListPairbyte[], byte[] fuzzyKeyDataCopy = new ArrayList(fuzzyKeysData.size()); +080 +081for (Pairbyte[], byte[] aFuzzyKeysData : fuzzyKeysData) { +082 if (aFuzzyKeysData.getFirst().length != aFuzzyKeysData.getSecond().length) { 083PairString, String readable = -084new Pair(Bytes.toStringBinary(p.getFirst()), Bytes.toStringBinary(p -085.getSecond())); -086throw new IllegalArgumentException("Fuzzy pair lengths do not match: " + readable); -087 } -088 // update mask ( 0 - -1 (0xff), 1 - 2) -089 p.setSecond(preprocessMask(p.getSecond())); -090 preprocessSearchKey(p); -091} -092this.fuzzyKeysData = fuzzyKeysData; -093this.tracker = new RowTracker(); -094 } -095 -096 private void preprocessSearchKey(Pairbyte[], byte[] p) { -097if (!UNSAFE_UNALIGNED) { -098 // do nothing -099 return; -100} -101byte[] key = p.getFirst(); -102byte[] mask = p.getSecond(); -103for (int i = 0; i mask.length; i++) { -104 // set non-fixed part of a search key to 0. -105 if (mask[i] == 2) { -106key[i] = 0; -107 } +084 new Pair(Bytes.toStringBinary(aFuzzyKeysData.getFirst()), Bytes.toStringBinary(aFuzzyKeysData.getSecond())); +085throw new IllegalArgumentException("Fuzzy pair lengths do not match: " + readable); +086 } +087 +088 Pairbyte[], byte[] p = new Pair(); +089 // create a copy of pair bytes so that they are not modified by the filter. +090 p.setFirst(Arrays.copyOf(aFuzzyKeysData.getFirst(), aFuzzyKeysData.getFirst().length)); +091 p.setSecond(Arrays.copyOf(aFuzzyKeysData.getSecond(), aFuzzyKeysData.getSecond().length)); +092 +093 // update mask ( 0 - -1 (0xff), 1 - 2) +094 p.setSecond(preprocessMask(p.getSecond())); +095 preprocessSearchKey(p); +096 +097 fuzzyKeyDataCopy.add(p); +098} +099this.fuzzyKeysData = fuzzyKeyDataCopy; +100this.tracker = new RowTracker(); +101 } +102 +103 +104 private void preprocessSearchKey(Pairbyte[], byte[] p) { +105if (!UNSAFE_UNALIGNED) { +106 // do nothing +107 return; 108} -109 } -110 -111 /** -112 * We need to preprocess mask array, as since we treat 2's as unfixed positions and -1 (0xff) as -113 * fixed positions -114 * @param mask -115 * @return mask array -116 */ -117 private byte[] preprocessMask(byte[] mask) { -118if (!UNSAFE_UNALIGNED) { -119 // do nothing -120 return mask; -121} -122if (isPreprocessedMask(mask)) return mask; -123for (int i = 0; i mask.length; i++) { -124 if (mask[i] == 0) { -125mask[i] = -1; // 0 - -1 -126 } else if (mask[i] == 1) { -127mask[i] = 2;// 1 - 2 -128 } +109byte[] key = p.getFirst(); +110byte[] mask = p.getSecond(); +111for (int i = 0; i mask.length; i++) { +112 // set non-fixed part of a search key to 0. +113 if (mask[i] == 2) { +114key[i] = 0; +115 } +116} +117 } +118 +119 /** +120 * We need to preprocess mask array, as since we treat 2's as unfixed positions and -1 (0xff) as +121 * fixed positions +122 * @param mask +123 * @return mask array +124 */ +125 private byte[] preprocessMask(byte[] mask) { +126if (!UNSAFE_UNALIGNED) { +127 // do nothing +128 return mask; 129} -130return mask; -131 } -132 -133 private boolean isPreprocessedMask(byte[] mask) { -134for (int i = 0; i mask.length; i++) { -135 if (mask[i] != -1 mask[i] != 2) { -136return false; -137 } -138} -139return true; -140 } -141 -142 @Override -143 public ReturnCode filterKeyValue(Cell c) { -144final int startIndex = lastFoundIndex = 0 ? lastFoundIndex : 0; -145final int size = fuzzyKeysData.size(); -146for (int i = startIndex; i size + startIndex; i++) { -147 final int index = i % size; -148 Pairbyte[], byte[] fuzzyData = fuzzyKeysData.get(index); -149 // This shift is idempotent - always end up with 0 and -1 as mask values. -150 for (int j = 0; j
[27/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/apidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html b/apidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html index a310286..e6ec67b 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.html @@ -84,552 +84,560 @@ 076 private RowTracker tracker; 077 078 public FuzzyRowFilter(ListPairbyte[], byte[] fuzzyKeysData) { -079Pairbyte[], byte[] p; -080for (int i = 0; i fuzzyKeysData.size(); i++) { -081 p = fuzzyKeysData.get(i); -082 if (p.getFirst().length != p.getSecond().length) { +079ListPairbyte[], byte[] fuzzyKeyDataCopy = new ArrayList(fuzzyKeysData.size()); +080 +081for (Pairbyte[], byte[] aFuzzyKeysData : fuzzyKeysData) { +082 if (aFuzzyKeysData.getFirst().length != aFuzzyKeysData.getSecond().length) { 083PairString, String readable = -084new Pair(Bytes.toStringBinary(p.getFirst()), Bytes.toStringBinary(p -085.getSecond())); -086throw new IllegalArgumentException("Fuzzy pair lengths do not match: " + readable); -087 } -088 // update mask ( 0 - -1 (0xff), 1 - 2) -089 p.setSecond(preprocessMask(p.getSecond())); -090 preprocessSearchKey(p); -091} -092this.fuzzyKeysData = fuzzyKeysData; -093this.tracker = new RowTracker(); -094 } -095 -096 private void preprocessSearchKey(Pairbyte[], byte[] p) { -097if (!UNSAFE_UNALIGNED) { -098 // do nothing -099 return; -100} -101byte[] key = p.getFirst(); -102byte[] mask = p.getSecond(); -103for (int i = 0; i mask.length; i++) { -104 // set non-fixed part of a search key to 0. -105 if (mask[i] == 2) { -106key[i] = 0; -107 } +084 new Pair(Bytes.toStringBinary(aFuzzyKeysData.getFirst()), Bytes.toStringBinary(aFuzzyKeysData.getSecond())); +085throw new IllegalArgumentException("Fuzzy pair lengths do not match: " + readable); +086 } +087 +088 Pairbyte[], byte[] p = new Pair(); +089 // create a copy of pair bytes so that they are not modified by the filter. +090 p.setFirst(Arrays.copyOf(aFuzzyKeysData.getFirst(), aFuzzyKeysData.getFirst().length)); +091 p.setSecond(Arrays.copyOf(aFuzzyKeysData.getSecond(), aFuzzyKeysData.getSecond().length)); +092 +093 // update mask ( 0 - -1 (0xff), 1 - 2) +094 p.setSecond(preprocessMask(p.getSecond())); +095 preprocessSearchKey(p); +096 +097 fuzzyKeyDataCopy.add(p); +098} +099this.fuzzyKeysData = fuzzyKeyDataCopy; +100this.tracker = new RowTracker(); +101 } +102 +103 +104 private void preprocessSearchKey(Pairbyte[], byte[] p) { +105if (!UNSAFE_UNALIGNED) { +106 // do nothing +107 return; 108} -109 } -110 -111 /** -112 * We need to preprocess mask array, as since we treat 2's as unfixed positions and -1 (0xff) as -113 * fixed positions -114 * @param mask -115 * @return mask array -116 */ -117 private byte[] preprocessMask(byte[] mask) { -118if (!UNSAFE_UNALIGNED) { -119 // do nothing -120 return mask; -121} -122if (isPreprocessedMask(mask)) return mask; -123for (int i = 0; i mask.length; i++) { -124 if (mask[i] == 0) { -125mask[i] = -1; // 0 - -1 -126 } else if (mask[i] == 1) { -127mask[i] = 2;// 1 - 2 -128 } +109byte[] key = p.getFirst(); +110byte[] mask = p.getSecond(); +111for (int i = 0; i mask.length; i++) { +112 // set non-fixed part of a search key to 0. +113 if (mask[i] == 2) { +114key[i] = 0; +115 } +116} +117 } +118 +119 /** +120 * We need to preprocess mask array, as since we treat 2's as unfixed positions and -1 (0xff) as +121 * fixed positions +122 * @param mask +123 * @return mask array +124 */ +125 private byte[] preprocessMask(byte[] mask) { +126if (!UNSAFE_UNALIGNED) { +127 // do nothing +128 return mask; 129} -130return mask; -131 } -132 -133 private boolean isPreprocessedMask(byte[] mask) { -134for (int i = 0; i mask.length; i++) { -135 if (mask[i] != -1 mask[i] != 2) { -136return false; -137 } -138} -139return true; -140 } -141 -142 @Override -143 public ReturnCode filterKeyValue(Cell c) { -144final int startIndex = lastFoundIndex = 0 ? lastFoundIndex : 0; -145final int size = fuzzyKeysData.size(); -146for (int i = startIndex; i size + startIndex; i++) { -147 final int index = i % size; -148 Pairbyte[], byte[] fuzzyData = fuzzyKeysData.get(index); -149 // This shift is idempotent - always end up with 0 and -1 as mask values. -150 for (int j = 0; j
[29/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/apache_hbase_reference_guide.pdf -- diff --git a/apache_hbase_reference_guide.pdf b/apache_hbase_reference_guide.pdf index e653e2b..5015387 100644 --- a/apache_hbase_reference_guide.pdf +++ b/apache_hbase_reference_guide.pdf @@ -5,16 +5,16 @@ /Author (Apache HBase Team) /Creator (Asciidoctor PDF 1.5.0.alpha.15, based on Prawn 2.2.2) /Producer (Apache HBase Team) -/ModDate (D:20170818144536+00'00') -/CreationDate (D:20170818144536+00'00') +/ModDate (D:20170819194146+00'00') +/CreationDate (D:20170819194146+00'00') >> endobj 2 0 obj << /Type /Catalog /Pages 3 0 R /Names 26 0 R -/Outlines 4311 0 R -/PageLabels 4519 0 R +/Outlines 4314 0 R +/PageLabels 4522 0 R /PageMode /UseOutlines /OpenAction [7 0 R /FitH 842.89] /ViewerPreferences << /DisplayDocTitle true @@ -23,8 +23,8 @@ endobj endobj 3 0 obj << /Type /Pages -/Count 662 -/Kids [7 0 R 12 0 R 14 0 R 16 0 R 18 0 R 20 0 R 22 0 R 24 0 R 44 0 R 47 0 R 50 0 R 54 0 R 63 0 R 66 0 R 69 0 R 71 0 R 76 0 R 80 0 R 83 0 R 89 0 R 91 0 R 94 0 R 96 0 R 103 0 R 109 0 R 114 0 R 116 0 R 130 0 R 133 0 R 142 0 R 150 0 R 160 0 R 169 0 R 180 0 R 184 0 R 186 0 R 190 0 R 201 0 R 208 0 R 217 0 R 225 0 R 231 0 R 239 0 R 248 0 R 261 0 R 270 0 R 277 0 R 286 0 R 295 0 R 302 0 R 309 0 R 317 0 R 323 0 R 331 0 R 338 0 R 347 0 R 356 0 R 367 0 R 378 0 R 386 0 R 393 0 R 401 0 R 408 0 R 416 0 R 425 0 R 434 0 R 441 0 R 449 0 R 462 0 R 469 0 R 477 0 R 484 0 R 493 0 R 501 0 R 508 0 R 513 0 R 518 0 R 523 0 R 539 0 R 549 0 R 554 0 R 568 0 R 574 0 R 579 0 R 581 0 R 583 0 R 588 0 R 596 0 R 602 0 R 607 0 R 613 0 R 624 0 R 637 0 R 656 0 R 671 0 R 682 0 R 684 0 R 686 0 R 694 0 R 706 0 R 715 0 R 725 0 R 731 0 R 734 0 R 738 0 R 743 0 R 746 0 R 749 0 R 751 0 R 754 0 R 758 0 R 760 0 R 765 0 R 769 0 R 774 0 R 779 0 R 782 0 R 788 0 R 790 0 R 794 0 R 803 0 R 805 0 R 808 0 R 811 0 R 814 0 R 817 0 R 831 0 R 838 0 R 847 0 R 858 0 R 864 0 R 876 0 R 880 0 R 883 0 R 887 0 R 890 0 R 895 0 R 904 0 R 912 0 R 916 0 R 920 0 R 925 0 R 929 0 R 931 0 R 946 0 R 957 0 R 963 0 R 969 0 R 972 0 R 980 0 R 988 0 R 992 0 R 998 0 R 1003 0 R 1005 0 R 1007 0 R 1009 0 R 1020 0 R 1028 0 R 1032 0 R 1039 0 R 1047 0 R 1055 0 R 1059 0 R 1065 0 R 1070 0 R 1078 0 R 1083 0 R 1088 0 R 1090 0 R 1096 0 R 1102 0 R 1104 0 R 0 R 1121 0 R 1125 0 R 1127 0 R 1131 0 R 1134 0 R 1139 0 R 1142 0 R 1154 0 R 1158 0 R 1164 0 R 1171 0 R 1176 0 R 1180 0 R 1184 0 R 1186 0 R 1189 0 R 1192 0 R 1195 0 R 1199 0 R 1203 0 R 1207 0 R 1212 0 R 1216 0 R 1219 0 R 1221 0 R 1233 0 R 1236 0 R 1244 0 R 1253 0 R 1259 0 R 1263 0 R 1265 0 R 1275 0 R 1278 0 R 1284 0 R 1292 0 R 1295 0 R 1302 0 R 1311 0 R 1313 0 R 1315 0 R 1325 0 R 1327 0 R 1329 0 R 1332 0 R 1334 0 R 1336 0 R 1338 0 R 1340 0 R 1343 0 R 1347 0 R 1352 0 R 1354 0 R 1356 0 R 1358 0 R 1363 0 R 1370 0 R 1375 0 R 1378 0 R 1380 0 R 1383 0 R 1387 0 R 1389 0 R 1392 0 R 1394 0 R 1396 0 R 1399 0 R 1404 0 R 1409 0 R 1417 0 R 1422 0 R 1436 0 R 1448 0 R 1451 0 R 1455 0 R 1469 0 R 1478 0 R 1492 0 R 1498 0 R 1506 0 R 1520 0 R 1534 0 R 1546 0 R 1551 0 R 1558 0 R 1569 0 R 1575 0 R 1581 0 R 1589 0 R 1592 0 R 1601 0 R 1608 0 R 1612 0 R 1625 0 R 1627 0 R 1633 0 R 1639 0 R 1643 0 R 1651 0 R 1660 0 R 1664 0 R 1666 0 R 1668 0 R 1681 0 R 1687 0 R 1695 0 R 1702 0 R 1716 0 R 1721 0 R 1730 0 R 1738 0 R 1744 0 R 1751 0 R 1755 0 R 1758 0 R 1760 0 R 1766 0 R 1772 0 R 1778 0 R 1782 0 R 1790 0 R 1795 0 R 1801 0 R 1806 0 R 1808 0 R 1817 0 R 1824 0 R 1830 0 R 1835 0 R 1839 0 R 1842 0 R 1847 0 R 1852 0 R 1859 0 R 1861 0 R 1863 0 R 1866 0 R 1874 0 R 1877 0 R 1884 0 R 1894 0 R 1897 0 R 1902 0 R 1904 0 R 1909 0 R 1912 0 R 1914 0 R 1919 0 R 1929 0 R 1931 0 R 1933 0 R 1935 0 R 1937 0 R 1940 0 R 1942 0 R 1944 0 R 1947 0 R 1949 0 R 1951 0 R 1955 0 R 1959 0 R 1968 0 R 1970 0 R 1972 0 R 1978 0 R 1980 0 R 1985 0 R 1987 0 R 1989 0 R 1996 0 R 2001 0 R 2005 0 R 2009 0 R 2013 0 R 2015 0 R 2017 0 R 2021 0 R 20 24 0 R 2026 0 R 2028 0 R 2032 0 R 2034 0 R 2037 0 R 2039 0 R 2041 0 R 2043 0 R 2050 0 R 2053 0 R 2058 0 R 2060 0 R 2062 0 R 2064 0 R 2066 0 R 2074 0 R 2085 0 R 2099 0 R 2110 0 R 2114 0 R 2119 0 R 2123 0 R 2126 0 R 2131 0 R 2137 0 R 2139 0 R 2143 0 R 2145 0 R 2147 0 R 2149 0 R 2153 0 R 2155 0 R 2168 0 R 2171 0 R 2179 0 R 2185 0 R 2197 0 R 2211 0 R 2225 0 R 2242 0 R 2246 0 R 2248 0 R 2252 0 R 2270 0 R 2276 0 R 2288 0 R 2292 0 R 2296 0 R 2305 0 R 2317 0 R 2323 0 R 2333 0 R 2346 0 R 2365 0 R 2374 0 R 2377 0 R 2386 0 R 2404 0 R 2411 0 R 2414 0 R 2419 0 R 2423 0 R 2426 0 R 2435 0 R 2444 0 R 2447 0 R 2449 0 R 2453 0 R 2468 0 R 2476 0 R 2481 0 R 2485 0 R 2489 0 R 2491 0 R 2493 0 R 2495 0 R 2500 0 R 2513 0 R 2523 0 R 2532 0 R 2541 0 R 2547 0 R 2558 0 R 2565 0 R 2571 0 R 2573 0 R 2583 0 R 2591 0 R 2601 0 R 2605 0 R 2616 0 R 2620 0 R 2630 0 R 2638 0 R 2646 0 R 2652 0 R 2656 0 R 2660 0 R 2664 0 R 2666 0 R 2672 0 R 2676 0 R 2680 0 R 2686 0 R 2692 0 R 2695 0 R 2701 0 R
[21/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.EmptyByteBufferCell.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.EmptyByteBufferCell.html b/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.EmptyByteBufferCell.html index 18d00df..3e734a6 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.EmptyByteBufferCell.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.EmptyByteBufferCell.html @@ -3143,7 +3143,60 @@ 3135 return Type.DeleteFamily.getCode(); 3136} 3137 } -3138} +3138 +3139 /** +3140 * @return An new cell is located following input cell. If both of type and timestamp are +3141 * minimum, the input cell will be returned directly. +3142 */ +3143 @InterfaceAudience.Private +3144 public static Cell createNextOnRowCol(Cell cell) { +3145long ts = cell.getTimestamp(); +3146byte type = cell.getTypeByte(); +3147if (type != Type.Minimum.getCode()) { +3148 type = KeyValue.Type.values()[KeyValue.Type.codeToType(type).ordinal() - 1].getCode(); +3149} else if (ts != HConstants.OLDEST_TIMESTAMP) { +3150 ts = ts - 1; +3151 type = Type.Maximum.getCode(); +3152} else { +3153 return cell; +3154} +3155return createNextOnRowCol(cell, ts, type); +3156 } +3157 +3158 private static Cell createNextOnRowCol(Cell cell, long ts, byte type) { +3159if (cell instanceof ByteBufferCell) { +3160 return new LastOnRowColByteBufferCell(((ByteBufferCell) cell).getRowByteBuffer(), +3161 ((ByteBufferCell) cell).getRowPosition(), cell.getRowLength(), +3162 ((ByteBufferCell) cell).getFamilyByteBuffer(), +3163 ((ByteBufferCell) cell).getFamilyPosition(), cell.getFamilyLength(), +3164 ((ByteBufferCell) cell).getQualifierByteBuffer(), +3165 ((ByteBufferCell) cell).getQualifierPosition(), cell.getQualifierLength()) { +3166@Override +3167public long getTimestamp() { +3168 return ts; +3169} +3170 +3171@Override +3172public byte getTypeByte() { +3173 return type; +3174} +3175 }; +3176} +3177return new LastOnRowColCell(cell.getRowArray(), cell.getRowOffset(), cell.getRowLength(), +3178cell.getFamilyArray(), cell.getFamilyOffset(), cell.getFamilyLength(), +3179cell.getQualifierArray(), cell.getQualifierOffset(), cell.getQualifierLength()) { +3180 @Override +3181 public long getTimestamp() { +3182return ts; +3183 } +3184 +3185 @Override +3186 public byte getTypeByte() { +3187return type; +3188 } +3189}; +3190 } +3191} http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.EmptyCell.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.EmptyCell.html b/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.EmptyCell.html index 18d00df..3e734a6 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.EmptyCell.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/CellUtil.EmptyCell.html @@ -3143,7 +3143,60 @@ 3135 return Type.DeleteFamily.getCode(); 3136} 3137 } -3138} +3138 +3139 /** +3140 * @return An new cell is located following input cell. If both of type and timestamp are +3141 * minimum, the input cell will be returned directly. +3142 */ +3143 @InterfaceAudience.Private +3144 public static Cell createNextOnRowCol(Cell cell) { +3145long ts = cell.getTimestamp(); +3146byte type = cell.getTypeByte(); +3147if (type != Type.Minimum.getCode()) { +3148 type = KeyValue.Type.values()[KeyValue.Type.codeToType(type).ordinal() - 1].getCode(); +3149} else if (ts != HConstants.OLDEST_TIMESTAMP) { +3150 ts = ts - 1; +3151 type = Type.Maximum.getCode(); +3152} else { +3153 return cell; +3154} +3155return createNextOnRowCol(cell, ts, type); +3156 } +3157 +3158 private static Cell createNextOnRowCol(Cell cell, long ts, byte type) { +3159if (cell instanceof ByteBufferCell) { +3160 return new LastOnRowColByteBufferCell(((ByteBufferCell) cell).getRowByteBuffer(), +3161 ((ByteBufferCell) cell).getRowPosition(), cell.getRowLength(), +3162 ((ByteBufferCell) cell).getFamilyByteBuffer(), +3163 ((ByteBufferCell) cell).getFamilyPosition(), cell.getFamilyLength(), +3164 ((ByteBufferCell) cell).getQualifierByteBuffer(), +3165 ((ByteBufferCell) cell).getQualifierPosition(), cell.getQualifierLength()) { +3166@Override +3167public long getTimestamp() { +3168 return ts; +3169} +3170 +3171
[06/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestFromClientSide3.WatiingForMultiMutationsObserver.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestFromClientSide3.WatiingForMultiMutationsObserver.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestFromClientSide3.WatiingForMultiMutationsObserver.html index 401b413..6bb0bed 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestFromClientSide3.WatiingForMultiMutationsObserver.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestFromClientSide3.WatiingForMultiMutationsObserver.html @@ -42,880 +42,981 @@ 034import org.apache.commons.logging.LogFactory; 035import org.apache.hadoop.conf.Configuration; 036import org.apache.hadoop.hbase.Cell; -037import org.apache.hadoop.hbase.Coprocessor; -038import org.apache.hadoop.hbase.coprocessor.ObserverContext; -039import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment; -040import org.apache.hadoop.hbase.coprocessor.RegionObserver; -041import org.apache.hadoop.hbase.regionserver.HRegionServer; -042import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress; -043import org.apache.hadoop.hbase.TableName; -044import org.apache.hadoop.hbase.regionserver.Region; -045import org.apache.hadoop.hbase.regionserver.RegionScanner; -046import org.apache.hadoop.hbase.HBaseTestingUtility; -047import org.apache.hadoop.hbase.HColumnDescriptor; -048import org.apache.hadoop.hbase.HConstants; -049import org.apache.hadoop.hbase.HRegionLocation; -050import org.apache.hadoop.hbase.HTableDescriptor; -051import org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint; -052import org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils; -053import org.apache.hadoop.hbase.ipc.ServerRpcController; -054import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos; -055import org.apache.hadoop.hbase.regionserver.HRegion; -056import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -057import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos; -058import org.apache.hadoop.hbase.testclassification.ClientTests; -059import org.apache.hadoop.hbase.testclassification.LargeTests; -060import org.apache.hadoop.hbase.util.Bytes; -061import org.apache.hadoop.hbase.util.Pair; -062import org.junit.After; -063import org.junit.AfterClass; -064import static org.junit.Assert.assertEquals; -065import static org.junit.Assert.assertFalse; -066import static org.junit.Assert.assertNull; -067import static org.junit.Assert.assertTrue; -068import static org.junit.Assert.fail; -069import org.junit.Before; -070import org.junit.BeforeClass; -071import org.junit.Rule; -072import org.junit.Test; -073import org.junit.experimental.categories.Category; -074import org.junit.rules.TestName; -075 -076@Category({LargeTests.class, ClientTests.class}) -077public class TestFromClientSide3 { -078 private static final Log LOG = LogFactory.getLog(TestFromClientSide3.class); -079 private final static HBaseTestingUtility TEST_UTIL -080= new HBaseTestingUtility(); -081 private static byte[] FAMILY = Bytes.toBytes("testFamily"); -082 private static Random random = new Random(); -083 private static int SLAVES = 3; -084 private static byte [] ROW = Bytes.toBytes("testRow"); -085 private static final byte[] ANOTHERROW = Bytes.toBytes("anotherrow"); -086 private static byte [] QUALIFIER = Bytes.toBytes("testQualifier"); -087 private static byte [] VALUE = Bytes.toBytes("testValue"); -088 private final static byte[] COL_QUAL = Bytes.toBytes("f1"); -089 private final static byte[] VAL_BYTES = Bytes.toBytes("v1"); -090 private final static byte[] ROW_BYTES = Bytes.toBytes("r1"); -091 -092 @Rule -093 public TestName name = new TestName(); -094 -095 /** -096 * @throws java.lang.Exception -097 */ -098 @BeforeClass -099 public static void setUpBeforeClass() throws Exception { -100TEST_UTIL.startMiniCluster(SLAVES); -101 } -102 -103 /** -104 * @throws java.lang.Exception -105 */ -106 @AfterClass -107 public static void tearDownAfterClass() throws Exception { -108TEST_UTIL.shutdownMiniCluster(); -109 } -110 -111 /** -112 * @throws java.lang.Exception -113 */ -114 @Before -115 public void setUp() throws Exception { -116// Nothing to do. -117 } -118 -119 /** -120 * @throws java.lang.Exception -121 */ -122 @After -123 public void tearDown() throws Exception { -124for (HTableDescriptor htd: TEST_UTIL.getAdmin().listTables()) { -125 LOG.info("Tear down, remove table=" + htd.getTableName()); -126 TEST_UTIL.deleteTable(htd.getTableName()); -127 } +037import org.apache.hadoop.hbase.CellUtil; +038import org.apache.hadoop.hbase.Coprocessor; +039import org.apache.hadoop.hbase.coprocessor.ObserverContext; +040import
[13/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/devapidocs/src-html/org/apache/hadoop/hbase/security/visibility/VisibilityScanDeleteTracker.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/security/visibility/VisibilityScanDeleteTracker.html b/devapidocs/src-html/org/apache/hadoop/hbase/security/visibility/VisibilityScanDeleteTracker.html index c884004..61c2f25 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/security/visibility/VisibilityScanDeleteTracker.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/security/visibility/VisibilityScanDeleteTracker.html @@ -28,311 +28,322 @@ 020 021import java.io.IOException; 022import java.util.ArrayList; -023import java.util.List; -024 -025import org.apache.commons.logging.Log; -026import org.apache.commons.logging.LogFactory; -027import org.apache.hadoop.hbase.classification.InterfaceAudience; -028import org.apache.hadoop.hbase.regionserver.querymatcher.ScanDeleteTracker; -029import org.apache.hadoop.hbase.Cell; -030import org.apache.hadoop.hbase.CellComparator; -031import org.apache.hadoop.hbase.CellUtil; -032import org.apache.hadoop.hbase.KeyValue; -033import org.apache.hadoop.hbase.KeyValue.Type; -034import org.apache.hadoop.hbase.Tag; -035import org.apache.hadoop.hbase.util.Bytes; -036import org.apache.hadoop.hbase.util.Pair; -037import org.apache.hadoop.hbase.util.Triple; -038 -039/** -040 * Similar to ScanDeletTracker but tracks the visibility expression also before -041 * deciding if a Cell can be considered deleted -042 */ -043@InterfaceAudience.Private -044public class VisibilityScanDeleteTracker extends ScanDeleteTracker { -045 -046 private static final Log LOG = LogFactory.getLog(VisibilityScanDeleteTracker.class); -047 -048 // Its better to track the visibility tags in delete based on each type. Create individual -049 // data structures for tracking each of them. This would ensure that there is no tracking based -050 // on time and also would handle all cases where deletefamily or deletecolumns is specified with -051 // Latest_timestamp. In such cases the ts in the delete marker and the masking -052 // put will not be same. So going with individual data structures for different delete -053 // type would solve this problem and also ensure that the combination of different type -054 // of deletes with diff ts would also work fine -055 // Track per TS -056 private ListTripleListTag, Byte, Long visibilityTagsDeleteFamily = new ArrayList(); -057 // Delete family version with different ts and different visibility expression could come. -058 // Need to track it per ts. -059 private ListTripleListTag, Byte, Long visibilityTagsDeleteFamilyVersion = new ArrayList(); -060 private ListPairListTag, Byte visibilityTagsDeleteColumns; -061 // Tracking as ListList is to handle same ts cell but different visibility tag. -062 // TODO : Need to handle puts with same ts but different vis tags. -063 private ListPairListTag, Byte visiblityTagsDeleteColumnVersion = new ArrayList(); -064 -065 public VisibilityScanDeleteTracker() { -066super(); -067 } -068 -069 @Override -070 public void add(Cell delCell) { -071//Cannot call super.add because need to find if the delete needs to be considered -072long timestamp = delCell.getTimestamp(); -073byte type = delCell.getTypeByte(); -074if (type == KeyValue.Type.DeleteFamily.getCode()) { -075 hasFamilyStamp = true; -076 boolean hasVisTag = extractDeleteCellVisTags(delCell, KeyValue.Type.DeleteFamily); -077 if (!hasVisTag timestamp familyStamp) { -078familyStamp = timestamp; -079 } -080 return; -081} else if (type == KeyValue.Type.DeleteFamilyVersion.getCode()) { -082 familyVersionStamps.add(timestamp); -083 extractDeleteCellVisTags(delCell, KeyValue.Type.DeleteFamilyVersion); -084 return; -085} -086// new column, or more general delete type -087if (deleteCell != null) { -088 if (!(CellUtil.matchingQualifier(delCell, deleteCell))) { -089// A case where there are deletes for a column qualifier but there are -090// no corresponding puts for them. Rare case. -091visibilityTagsDeleteColumns = null; -092visiblityTagsDeleteColumnVersion = null; -093 } else if (type == KeyValue.Type.Delete.getCode() (deleteTimestamp != timestamp)) { -094// there is a timestamp change which means we could clear the list -095// when ts is same and the vis tags are different we need to collect -096// them all. Interesting part is that in the normal case of puts if -097// there are 2 cells with same ts and diff vis tags only one of them is -098// returned. Handling with a single ListTag would mean that only one -099// of the cell would be considered. Doing this as a precaution. -100// Rare
hbase-site git commit: INFRA-10751 Empty commit
Repository: hbase-site Updated Branches: refs/heads/asf-site 495ddb865 -> 16b7dcb03 INFRA-10751 Empty commit Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/16b7dcb0 Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/16b7dcb0 Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/16b7dcb0 Branch: refs/heads/asf-site Commit: 16b7dcb03003c82f0e59bb372cc344d14be3e8b9 Parents: 495ddb8 Author: jenkinsAuthored: Sat Aug 19 20:02:36 2017 + Committer: jenkins Committed: Sat Aug 19 20:02:36 2017 + -- --
[30/30] hbase-site git commit: Published site at .
Published site at . Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/495ddb86 Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/495ddb86 Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/495ddb86 Branch: refs/heads/asf-site Commit: 495ddb865369291c7709f779c134fcc5b3e37524 Parents: 2d30087 Author: jenkinsAuthored: Sat Aug 19 20:02:17 2017 + Committer: jenkins Committed: Sat Aug 19 20:02:17 2017 + -- acid-semantics.html | 4 +- apache_hbase_reference_guide.pdf| 16558 + apidocs/index-all.html | 2 + .../org/apache/hadoop/hbase/class-use/Cell.html | 142 +- .../hadoop/hbase/filter/FuzzyRowFilter.html |12 +- .../org/apache/hadoop/hbase/CellUtil.html |55 +- .../hadoop/hbase/filter/FuzzyRowFilter.html | 1084 +- book.html |66 +- bulk-loads.html | 4 +- checkstyle-aggregate.html | 12970 +++-- checkstyle.rss | 8 +- coc.html| 4 +- cygwin.html | 4 +- dependencies.html | 4 +- dependency-convergence.html | 4 +- dependency-info.html| 4 +- dependency-management.html | 4 +- devapidocs/constant-values.html | 6 +- devapidocs/index-all.html | 8 + .../org/apache/hadoop/hbase/CellUtil.html | 186 +- .../hadoop/hbase/backup/package-tree.html | 2 +- .../org/apache/hadoop/hbase/class-use/Cell.html |20 + .../org/apache/hadoop/hbase/class-use/Tag.html |14 +- .../class-use/InterfaceAudience.Private.html|38 +- .../hbase/classification/package-tree.html | 6 +- .../hadoop/hbase/client/package-tree.html |24 +- .../hadoop/hbase/executor/package-tree.html | 2 +- .../hbase/filter/FuzzyRowFilter.Order.html |22 +- .../hbase/filter/FuzzyRowFilter.RowTracker.html |16 +- .../filter/FuzzyRowFilter.SatisfiesCode.html|12 +- .../hadoop/hbase/filter/FuzzyRowFilter.html |36 +- .../hadoop/hbase/filter/package-tree.html |10 +- .../hadoop/hbase/io/hfile/package-tree.html | 6 +- .../hadoop/hbase/mapreduce/package-tree.html| 2 +- .../hbase/master/balancer/package-tree.html | 2 +- .../hadoop/hbase/master/package-tree.html | 6 +- .../hbase/master/procedure/package-tree.html| 2 +- .../org/apache/hadoop/hbase/package-tree.html |14 +- .../hadoop/hbase/procedure2/package-tree.html | 6 +- .../hadoop/hbase/quotas/package-tree.html |10 +- .../hadoop/hbase/regionserver/package-tree.html |18 +- .../querymatcher/ScanQueryMatcher.html |16 +- .../regionserver/querymatcher/package-tree.html | 4 +- .../hbase/regionserver/wal/package-tree.html| 2 +- .../hadoop/hbase/rest/model/package-tree.html | 2 +- .../hbase/security/access/package-tree.html | 2 +- .../DefaultVisibilityLabelServiceImpl.html |14 +- ...ibilityController.VisibilityReplication.html |14 +- .../visibility/VisibilityController.html| 2 +- .../visibility/VisibilityScanDeleteTracker.html |46 +- .../apache/hadoop/hbase/util/package-tree.html |10 +- .../apache/hadoop/hbase/wal/package-tree.html | 2 +- .../hbase/CellUtil.EmptyByteBufferCell.html |55 +- .../apache/hadoop/hbase/CellUtil.EmptyCell.html |55 +- .../CellUtil.FirstOnRowByteBufferCell.html |55 +- .../hadoop/hbase/CellUtil.FirstOnRowCell.html |55 +- .../CellUtil.FirstOnRowColByteBufferCell.html |55 +- .../hbase/CellUtil.FirstOnRowColCell.html |55 +- .../CellUtil.FirstOnRowColTSByteBufferCell.html |55 +- .../hbase/CellUtil.FirstOnRowColTSCell.html |55 +- .../CellUtil.FirstOnRowDeleteFamilyCell.html|55 +- .../hbase/CellUtil.LastOnRowByteBufferCell.html |55 +- .../hadoop/hbase/CellUtil.LastOnRowCell.html|55 +- .../CellUtil.LastOnRowColByteBufferCell.html|55 +- .../hadoop/hbase/CellUtil.LastOnRowColCell.html |55 +- .../CellUtil.TagRewriteByteBufferCell.html |55 +- .../hadoop/hbase/CellUtil.TagRewriteCell.html |55 +- ...llUtil.ValueAndTagRewriteByteBufferCell.html |55 +- .../hbase/CellUtil.ValueAndTagRewriteCell.html |55 +- .../org/apache/hadoop/hbase/CellUtil.html |55 +- .../org/apache/hadoop/hbase/Version.html| 6 +- .../hbase/filter/FuzzyRowFilter.Order.html | 1084 +-
[08/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/testdevapidocs/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.html b/testdevapidocs/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.html index db3cffc..e9224de 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.html +++ b/testdevapidocs/org/apache/hadoop/hbase/security/visibility/TestVisibilityLabelsWithDeletes.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":9,"i1":9,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":9,"i11":10,"i12":9,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10}; +var methods = {"i0":9,"i1":9,"i2":9,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":9,"i12":10,"i13":9,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10}; var tabs = {65535:["t0","All Methods"],1:["t1","Static Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -50,7 +50,7 @@ var activeTableTab = "activeTableTab"; PrevClass -NextClass +NextClass Frames @@ -74,7 +74,7 @@ var activeTableTab = "activeTableTab"; Summary: -Nested| +Nested| Field| Constr| Method @@ -113,7 +113,7 @@ var activeTableTab = "activeTableTab"; -public class TestVisibilityLabelsWithDeletes +public class TestVisibilityLabelsWithDeletes extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object Tests visibility labels with deletes @@ -122,6 +122,25 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? + + + + + +Nested Class Summary + +Nested Classes + +Modifier and Type +Class and Description + + +private static class +TestVisibilityLabelsWithDeletes.DeleteMark + + + + @@ -147,58 +166,62 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? fam +private static org.apache.commons.logging.Log +LOG + + private static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String PRIVATE - + private static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String PUBLIC - + protected static byte[] qual - + private static byte[] qual1 - + private static byte[] qual2 - + private static byte[] row1 - + private static byte[] row2 - + private static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String SECRET - + static org.apache.hadoop.hbase.security.User SUPERUSER - + org.junit.rules.TestName TEST_NAME - + static HBaseTestingUtility TEST_UTIL - + private static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String TOPSECRET - + protected static byte[] value - + private static byte[] value1 @@ -235,231 +258,257 @@ extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html? Method and Description +private static org.apache.hadoop.hbase.client.Delete +addDeleteMark(org.apache.hadoop.hbase.client.Deleted, + TestVisibilityLabelsWithDeletes.DeleteMarkmark, + longnow) + + static void addLabels() - + static Thttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListT createList(T...ts) - + protected org.apache.hadoop.hbase.client.Table createTable(org.apache.hadoop.hbase.HColumnDescriptorfam) - + org.apache.hadoop.hbase.client.Table createTableAndWriteDataWithLabels(long[]timestamp, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String...labelExps) - + org.apache.hadoop.hbase.client.Table
[19/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.Order.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.Order.html b/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.Order.html index a310286..e6ec67b 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.Order.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.Order.html @@ -84,552 +84,560 @@ 076 private RowTracker tracker; 077 078 public FuzzyRowFilter(ListPairbyte[], byte[] fuzzyKeysData) { -079Pairbyte[], byte[] p; -080for (int i = 0; i fuzzyKeysData.size(); i++) { -081 p = fuzzyKeysData.get(i); -082 if (p.getFirst().length != p.getSecond().length) { +079ListPairbyte[], byte[] fuzzyKeyDataCopy = new ArrayList(fuzzyKeysData.size()); +080 +081for (Pairbyte[], byte[] aFuzzyKeysData : fuzzyKeysData) { +082 if (aFuzzyKeysData.getFirst().length != aFuzzyKeysData.getSecond().length) { 083PairString, String readable = -084new Pair(Bytes.toStringBinary(p.getFirst()), Bytes.toStringBinary(p -085.getSecond())); -086throw new IllegalArgumentException("Fuzzy pair lengths do not match: " + readable); -087 } -088 // update mask ( 0 - -1 (0xff), 1 - 2) -089 p.setSecond(preprocessMask(p.getSecond())); -090 preprocessSearchKey(p); -091} -092this.fuzzyKeysData = fuzzyKeysData; -093this.tracker = new RowTracker(); -094 } -095 -096 private void preprocessSearchKey(Pairbyte[], byte[] p) { -097if (!UNSAFE_UNALIGNED) { -098 // do nothing -099 return; -100} -101byte[] key = p.getFirst(); -102byte[] mask = p.getSecond(); -103for (int i = 0; i mask.length; i++) { -104 // set non-fixed part of a search key to 0. -105 if (mask[i] == 2) { -106key[i] = 0; -107 } +084 new Pair(Bytes.toStringBinary(aFuzzyKeysData.getFirst()), Bytes.toStringBinary(aFuzzyKeysData.getSecond())); +085throw new IllegalArgumentException("Fuzzy pair lengths do not match: " + readable); +086 } +087 +088 Pairbyte[], byte[] p = new Pair(); +089 // create a copy of pair bytes so that they are not modified by the filter. +090 p.setFirst(Arrays.copyOf(aFuzzyKeysData.getFirst(), aFuzzyKeysData.getFirst().length)); +091 p.setSecond(Arrays.copyOf(aFuzzyKeysData.getSecond(), aFuzzyKeysData.getSecond().length)); +092 +093 // update mask ( 0 - -1 (0xff), 1 - 2) +094 p.setSecond(preprocessMask(p.getSecond())); +095 preprocessSearchKey(p); +096 +097 fuzzyKeyDataCopy.add(p); +098} +099this.fuzzyKeysData = fuzzyKeyDataCopy; +100this.tracker = new RowTracker(); +101 } +102 +103 +104 private void preprocessSearchKey(Pairbyte[], byte[] p) { +105if (!UNSAFE_UNALIGNED) { +106 // do nothing +107 return; 108} -109 } -110 -111 /** -112 * We need to preprocess mask array, as since we treat 2's as unfixed positions and -1 (0xff) as -113 * fixed positions -114 * @param mask -115 * @return mask array -116 */ -117 private byte[] preprocessMask(byte[] mask) { -118if (!UNSAFE_UNALIGNED) { -119 // do nothing -120 return mask; -121} -122if (isPreprocessedMask(mask)) return mask; -123for (int i = 0; i mask.length; i++) { -124 if (mask[i] == 0) { -125mask[i] = -1; // 0 - -1 -126 } else if (mask[i] == 1) { -127mask[i] = 2;// 1 - 2 -128 } +109byte[] key = p.getFirst(); +110byte[] mask = p.getSecond(); +111for (int i = 0; i mask.length; i++) { +112 // set non-fixed part of a search key to 0. +113 if (mask[i] == 2) { +114key[i] = 0; +115 } +116} +117 } +118 +119 /** +120 * We need to preprocess mask array, as since we treat 2's as unfixed positions and -1 (0xff) as +121 * fixed positions +122 * @param mask +123 * @return mask array +124 */ +125 private byte[] preprocessMask(byte[] mask) { +126if (!UNSAFE_UNALIGNED) { +127 // do nothing +128 return mask; 129} -130return mask; -131 } -132 -133 private boolean isPreprocessedMask(byte[] mask) { -134for (int i = 0; i mask.length; i++) { -135 if (mask[i] != -1 mask[i] != 2) { -136return false; -137 } -138} -139return true; -140 } -141 -142 @Override -143 public ReturnCode filterKeyValue(Cell c) { -144final int startIndex = lastFoundIndex = 0 ? lastFoundIndex : 0; -145final int size = fuzzyKeysData.size(); -146for (int i = startIndex; i size + startIndex; i++) { -147 final int index = i % size; -148 Pairbyte[], byte[] fuzzyData = fuzzyKeysData.get(index); -149 // This shift is idempotent - always end up with 0 and -1 as mask values. -150
[18/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.RowTracker.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.RowTracker.html b/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.RowTracker.html index a310286..e6ec67b 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.RowTracker.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/filter/FuzzyRowFilter.RowTracker.html @@ -84,552 +84,560 @@ 076 private RowTracker tracker; 077 078 public FuzzyRowFilter(ListPairbyte[], byte[] fuzzyKeysData) { -079Pairbyte[], byte[] p; -080for (int i = 0; i fuzzyKeysData.size(); i++) { -081 p = fuzzyKeysData.get(i); -082 if (p.getFirst().length != p.getSecond().length) { +079ListPairbyte[], byte[] fuzzyKeyDataCopy = new ArrayList(fuzzyKeysData.size()); +080 +081for (Pairbyte[], byte[] aFuzzyKeysData : fuzzyKeysData) { +082 if (aFuzzyKeysData.getFirst().length != aFuzzyKeysData.getSecond().length) { 083PairString, String readable = -084new Pair(Bytes.toStringBinary(p.getFirst()), Bytes.toStringBinary(p -085.getSecond())); -086throw new IllegalArgumentException("Fuzzy pair lengths do not match: " + readable); -087 } -088 // update mask ( 0 - -1 (0xff), 1 - 2) -089 p.setSecond(preprocessMask(p.getSecond())); -090 preprocessSearchKey(p); -091} -092this.fuzzyKeysData = fuzzyKeysData; -093this.tracker = new RowTracker(); -094 } -095 -096 private void preprocessSearchKey(Pairbyte[], byte[] p) { -097if (!UNSAFE_UNALIGNED) { -098 // do nothing -099 return; -100} -101byte[] key = p.getFirst(); -102byte[] mask = p.getSecond(); -103for (int i = 0; i mask.length; i++) { -104 // set non-fixed part of a search key to 0. -105 if (mask[i] == 2) { -106key[i] = 0; -107 } +084 new Pair(Bytes.toStringBinary(aFuzzyKeysData.getFirst()), Bytes.toStringBinary(aFuzzyKeysData.getSecond())); +085throw new IllegalArgumentException("Fuzzy pair lengths do not match: " + readable); +086 } +087 +088 Pairbyte[], byte[] p = new Pair(); +089 // create a copy of pair bytes so that they are not modified by the filter. +090 p.setFirst(Arrays.copyOf(aFuzzyKeysData.getFirst(), aFuzzyKeysData.getFirst().length)); +091 p.setSecond(Arrays.copyOf(aFuzzyKeysData.getSecond(), aFuzzyKeysData.getSecond().length)); +092 +093 // update mask ( 0 - -1 (0xff), 1 - 2) +094 p.setSecond(preprocessMask(p.getSecond())); +095 preprocessSearchKey(p); +096 +097 fuzzyKeyDataCopy.add(p); +098} +099this.fuzzyKeysData = fuzzyKeyDataCopy; +100this.tracker = new RowTracker(); +101 } +102 +103 +104 private void preprocessSearchKey(Pairbyte[], byte[] p) { +105if (!UNSAFE_UNALIGNED) { +106 // do nothing +107 return; 108} -109 } -110 -111 /** -112 * We need to preprocess mask array, as since we treat 2's as unfixed positions and -1 (0xff) as -113 * fixed positions -114 * @param mask -115 * @return mask array -116 */ -117 private byte[] preprocessMask(byte[] mask) { -118if (!UNSAFE_UNALIGNED) { -119 // do nothing -120 return mask; -121} -122if (isPreprocessedMask(mask)) return mask; -123for (int i = 0; i mask.length; i++) { -124 if (mask[i] == 0) { -125mask[i] = -1; // 0 - -1 -126 } else if (mask[i] == 1) { -127mask[i] = 2;// 1 - 2 -128 } +109byte[] key = p.getFirst(); +110byte[] mask = p.getSecond(); +111for (int i = 0; i mask.length; i++) { +112 // set non-fixed part of a search key to 0. +113 if (mask[i] == 2) { +114key[i] = 0; +115 } +116} +117 } +118 +119 /** +120 * We need to preprocess mask array, as since we treat 2's as unfixed positions and -1 (0xff) as +121 * fixed positions +122 * @param mask +123 * @return mask array +124 */ +125 private byte[] preprocessMask(byte[] mask) { +126if (!UNSAFE_UNALIGNED) { +127 // do nothing +128 return mask; 129} -130return mask; -131 } -132 -133 private boolean isPreprocessedMask(byte[] mask) { -134for (int i = 0; i mask.length; i++) { -135 if (mask[i] != -1 mask[i] != 2) { -136return false; -137 } -138} -139return true; -140 } -141 -142 @Override -143 public ReturnCode filterKeyValue(Cell c) { -144final int startIndex = lastFoundIndex = 0 ? lastFoundIndex : 0; -145final int size = fuzzyKeysData.size(); -146for (int i = startIndex; i size + startIndex; i++) { -147 final int index = i % size; -148 Pairbyte[], byte[] fuzzyData = fuzzyKeysData.get(index); -149 // This shift is idempotent - always end up with 0
[11/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/hbase-spark/dependency-convergence.html -- diff --git a/hbase-spark/dependency-convergence.html b/hbase-spark/dependency-convergence.html index 94ce635..b8e1164 100644 --- a/hbase-spark/dependency-convergence.html +++ b/hbase-spark/dependency-convergence.html @@ -7,7 +7,7 @@ - + Apache HBase - Spark Reactor Dependency Convergence @@ -571,7 +571,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-08-18 + Last Published: 2017-08-19 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/hbase-spark/dependency-info.html -- diff --git a/hbase-spark/dependency-info.html b/hbase-spark/dependency-info.html index 1025a52..05eb5c8 100644 --- a/hbase-spark/dependency-info.html +++ b/hbase-spark/dependency-info.html @@ -7,7 +7,7 @@ - + Apache HBase - Spark Dependency Information @@ -147,7 +147,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-08-18 + Last Published: 2017-08-19 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/hbase-spark/dependency-management.html -- diff --git a/hbase-spark/dependency-management.html b/hbase-spark/dependency-management.html index f556e98..dd21465 100644 --- a/hbase-spark/dependency-management.html +++ b/hbase-spark/dependency-management.html @@ -7,7 +7,7 @@ - + Apache HBase - Spark Project Dependency Management @@ -736,7 +736,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-08-18 + Last Published: 2017-08-19 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/hbase-spark/index.html -- diff --git a/hbase-spark/index.html b/hbase-spark/index.html index 99bafdb..ff6ab1d 100644 --- a/hbase-spark/index.html +++ b/hbase-spark/index.html @@ -7,7 +7,7 @@ - + Apache HBase - Spark About @@ -122,7 +122,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-08-18 + Last Published: 2017-08-19 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/hbase-spark/integration.html -- diff --git a/hbase-spark/integration.html b/hbase-spark/integration.html index 4b905a6..711d75a 100644 --- a/hbase-spark/integration.html +++ b/hbase-spark/integration.html @@ -7,7 +7,7 @@ - + Apache HBase - Spark CI Management @@ -126,7 +126,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-08-18 + Last Published: 2017-08-19 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/hbase-spark/issue-tracking.html -- diff --git a/hbase-spark/issue-tracking.html b/hbase-spark/issue-tracking.html index eab1401..60d01b1 100644 --- a/hbase-spark/issue-tracking.html +++ b/hbase-spark/issue-tracking.html @@ -7,7 +7,7 @@ - + Apache HBase - Spark Issue Management @@ -123,7 +123,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-08-18 + Last Published: 2017-08-19 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/hbase-spark/license.html -- diff --git a/hbase-spark/license.html b/hbase-spark/license.html index 05da248..8aacbfd 100644 --- a/hbase-spark/license.html +++ b/hbase-spark/license.html @@ -7,7 +7,7 @@ - + Apache HBase - Spark Project Licenses @@ -326,7 +326,7 @@
[04/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestFromClientSide3.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestFromClientSide3.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestFromClientSide3.html index 401b413..6bb0bed 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestFromClientSide3.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/client/TestFromClientSide3.html @@ -42,880 +42,981 @@ 034import org.apache.commons.logging.LogFactory; 035import org.apache.hadoop.conf.Configuration; 036import org.apache.hadoop.hbase.Cell; -037import org.apache.hadoop.hbase.Coprocessor; -038import org.apache.hadoop.hbase.coprocessor.ObserverContext; -039import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment; -040import org.apache.hadoop.hbase.coprocessor.RegionObserver; -041import org.apache.hadoop.hbase.regionserver.HRegionServer; -042import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress; -043import org.apache.hadoop.hbase.TableName; -044import org.apache.hadoop.hbase.regionserver.Region; -045import org.apache.hadoop.hbase.regionserver.RegionScanner; -046import org.apache.hadoop.hbase.HBaseTestingUtility; -047import org.apache.hadoop.hbase.HColumnDescriptor; -048import org.apache.hadoop.hbase.HConstants; -049import org.apache.hadoop.hbase.HRegionLocation; -050import org.apache.hadoop.hbase.HTableDescriptor; -051import org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint; -052import org.apache.hadoop.hbase.ipc.CoprocessorRpcUtils; -053import org.apache.hadoop.hbase.ipc.ServerRpcController; -054import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos; -055import org.apache.hadoop.hbase.regionserver.HRegion; -056import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; -057import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos; -058import org.apache.hadoop.hbase.testclassification.ClientTests; -059import org.apache.hadoop.hbase.testclassification.LargeTests; -060import org.apache.hadoop.hbase.util.Bytes; -061import org.apache.hadoop.hbase.util.Pair; -062import org.junit.After; -063import org.junit.AfterClass; -064import static org.junit.Assert.assertEquals; -065import static org.junit.Assert.assertFalse; -066import static org.junit.Assert.assertNull; -067import static org.junit.Assert.assertTrue; -068import static org.junit.Assert.fail; -069import org.junit.Before; -070import org.junit.BeforeClass; -071import org.junit.Rule; -072import org.junit.Test; -073import org.junit.experimental.categories.Category; -074import org.junit.rules.TestName; -075 -076@Category({LargeTests.class, ClientTests.class}) -077public class TestFromClientSide3 { -078 private static final Log LOG = LogFactory.getLog(TestFromClientSide3.class); -079 private final static HBaseTestingUtility TEST_UTIL -080= new HBaseTestingUtility(); -081 private static byte[] FAMILY = Bytes.toBytes("testFamily"); -082 private static Random random = new Random(); -083 private static int SLAVES = 3; -084 private static byte [] ROW = Bytes.toBytes("testRow"); -085 private static final byte[] ANOTHERROW = Bytes.toBytes("anotherrow"); -086 private static byte [] QUALIFIER = Bytes.toBytes("testQualifier"); -087 private static byte [] VALUE = Bytes.toBytes("testValue"); -088 private final static byte[] COL_QUAL = Bytes.toBytes("f1"); -089 private final static byte[] VAL_BYTES = Bytes.toBytes("v1"); -090 private final static byte[] ROW_BYTES = Bytes.toBytes("r1"); -091 -092 @Rule -093 public TestName name = new TestName(); -094 -095 /** -096 * @throws java.lang.Exception -097 */ -098 @BeforeClass -099 public static void setUpBeforeClass() throws Exception { -100TEST_UTIL.startMiniCluster(SLAVES); -101 } -102 -103 /** -104 * @throws java.lang.Exception -105 */ -106 @AfterClass -107 public static void tearDownAfterClass() throws Exception { -108TEST_UTIL.shutdownMiniCluster(); -109 } -110 -111 /** -112 * @throws java.lang.Exception -113 */ -114 @Before -115 public void setUp() throws Exception { -116// Nothing to do. -117 } -118 -119 /** -120 * @throws java.lang.Exception -121 */ -122 @After -123 public void tearDown() throws Exception { -124for (HTableDescriptor htd: TEST_UTIL.getAdmin().listTables()) { -125 LOG.info("Tear down, remove table=" + htd.getTableName()); -126 TEST_UTIL.deleteTable(htd.getTableName()); -127 } +037import org.apache.hadoop.hbase.CellUtil; +038import org.apache.hadoop.hbase.Coprocessor; +039import org.apache.hadoop.hbase.coprocessor.ObserverContext; +040import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment; +041import org.apache.hadoop.hbase.coprocessor.RegionObserver; +042import
[23/30] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/devapidocs/org/apache/hadoop/hbase/classification/class-use/InterfaceAudience.Private.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/classification/class-use/InterfaceAudience.Private.html b/devapidocs/org/apache/hadoop/hbase/classification/class-use/InterfaceAudience.Private.html index 7b2e58c..b5b8b17 100644 --- a/devapidocs/org/apache/hadoop/hbase/classification/class-use/InterfaceAudience.Private.html +++ b/devapidocs/org/apache/hadoop/hbase/classification/class-use/InterfaceAudience.Private.html @@ -1027,33 +1027,37 @@ service. +static Cell +CellUtil.createNextOnRowCol(Cellcell) + + long ProcedureInfo.getClientAckTime() - + static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String AuthUtil.getGroupName(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringaclKey) Returns the actual name for a group principal (stripped of the group prefix). - + static byte CellUtil.getQualifierByte(Cellcell, intindex) - + http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListRegionState ClusterStatus.getRegionsInTransition() - + static byte CellUtil.getRowByte(Cellcell, intindex) misc - + KeyValue.KVComparator TableName.getRowComparator() Deprecated. @@ -1062,67 +1066,67 @@ service. - + static ServerName MetaTableAccessor.getServerName(Resultr, intreplicaId) Returns a ServerName from catalog table Result. - + boolean ProcedureInfo.hasClientAckTime() - + boolean ChoreService.isChoreScheduled(ScheduledChorechore) - + static boolean AuthUtil.isGroupPrincipal(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) Returns whether or not the given name should be interpreted as a group principal. - + static boolean ProcedureInfo.isProcedureOwner(ProcedureInfoprocInfo, Useruser) Check if the user is this procedure's owner - + org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ServerLoad ServerLoad.obtainServerLoadPB() - + void ChoreService.onChoreMissedStartTime(ScheduledChorechore) - + void ProcedureInfo.setClientAckTime(longtimestamp) - + static void CellUtil.setSequenceId(Cellcell, longseqId) Sets the given seqId to the cell. - + static http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String AuthUtil.toGroupEntry(http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">Stringname) Returns the group entry with the group prefix for a group principal. - + http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String ScheduledChore.toString() A summation of this chore in human readable format. - + boolean ChoreService.triggerNow(ScheduledChorechore) http://git-wip-us.apache.org/repos/asf/hbase-site/blob/495ddb86/devapidocs/org/apache/hadoop/hbase/classification/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/classification/package-tree.html b/devapidocs/org/apache/hadoop/hbase/classification/package-tree.html index b8df865..f28d704 100644 --- a/devapidocs/org/apache/hadoop/hbase/classification/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/classification/package-tree.html @@ -88,12 +88,12 @@ Annotation Type Hierarchy +org.apache.hadoop.hbase.classification.InterfaceAudience.Private (implements java.lang.annotation.http://docs.oracle.com/javase/8/docs/api/java/lang/annotation/Annotation.html?is-external=true; title="class or interface in java.lang.annotation">Annotation) +org.apache.hadoop.hbase.classification.InterfaceStability.Unstable (implements java.lang.annotation.http://docs.oracle.com/javase/8/docs/api/java/lang/annotation/Annotation.html?is-external=true; title="class or interface in java.lang.annotation">Annotation) +org.apache.hadoop.hbase.classification.InterfaceAudience.Public (implements java.lang.annotation.http://docs.oracle.com/javase/8/docs/api/java/lang/annotation/Annotation.html?is-external=true; title="class or interface in java.lang.annotation">Annotation) org.apache.hadoop.hbase.classification.InterfaceStability.Evolving (implements java.lang.annotation.http://docs.oracle.com/javase/8/docs/api/java/lang/annotation/Annotation.html?is-external=true; title="class or interface in java.lang.annotation">Annotation)