[1/4] hbase git commit: HBASE-20212 Make all Public classes have InterfaceAudience category
Repository: hbase Updated Branches: refs/heads/master 68b2f5502 -> a6eeb26cc http://git-wip-us.apache.org/repos/asf/hbase/blob/a6eeb26c/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/example/hbasecontext/JavaHBaseBulkGetExample.java -- diff --git a/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/example/hbasecontext/JavaHBaseBulkGetExample.java b/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/example/hbasecontext/JavaHBaseBulkGetExample.java index cb9e0c7..b5143de 100644 --- a/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/example/hbasecontext/JavaHBaseBulkGetExample.java +++ b/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/example/hbasecontext/JavaHBaseBulkGetExample.java @@ -19,7 +19,6 @@ package org.apache.hadoop.hbase.spark.example.hbasecontext; import java.util.ArrayList; import java.util.Iterator; import java.util.List; - import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.HBaseConfiguration; @@ -32,11 +31,13 @@ import org.apache.spark.SparkConf; import org.apache.spark.api.java.JavaRDD; import org.apache.spark.api.java.JavaSparkContext; import org.apache.spark.api.java.function.Function; +import org.apache.yetus.audience.InterfaceAudience; /** * This is a simple example of getting records in HBase * with the bulkGet function. */ +@InterfaceAudience.Private final public class JavaHBaseBulkGetExample { private JavaHBaseBulkGetExample() {} http://git-wip-us.apache.org/repos/asf/hbase/blob/a6eeb26c/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/example/hbasecontext/JavaHBaseBulkLoadExample.java -- diff --git a/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/example/hbasecontext/JavaHBaseBulkLoadExample.java b/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/example/hbasecontext/JavaHBaseBulkLoadExample.java index f0f3e79..6738059 100644 --- a/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/example/hbasecontext/JavaHBaseBulkLoadExample.java +++ b/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/example/hbasecontext/JavaHBaseBulkLoadExample.java @@ -19,7 +19,6 @@ package org.apache.hadoop.hbase.spark.example.hbasecontext; import java.util.ArrayList; import java.util.HashMap; import java.util.List; - import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.HConstants; @@ -33,6 +32,7 @@ import org.apache.spark.SparkConf; import org.apache.spark.api.java.JavaRDD; import org.apache.spark.api.java.JavaSparkContext; import org.apache.spark.api.java.function.Function; +import org.apache.yetus.audience.InterfaceAudience; /** * Run this example using command below: @@ -45,6 +45,7 @@ import org.apache.spark.api.java.function.Function; * 'hbase org.apache.hadoop.hbase.tool.LoadIncrementalHFiles' to load the HFiles into table to * verify this example. */ +@InterfaceAudience.Private final public class JavaHBaseBulkLoadExample { private JavaHBaseBulkLoadExample() {} http://git-wip-us.apache.org/repos/asf/hbase/blob/a6eeb26c/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/example/hbasecontext/JavaHBaseBulkPutExample.java -- diff --git a/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/example/hbasecontext/JavaHBaseBulkPutExample.java b/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/example/hbasecontext/JavaHBaseBulkPutExample.java index 5821c19..4a80b96 100644 --- a/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/example/hbasecontext/JavaHBaseBulkPutExample.java +++ b/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/example/hbasecontext/JavaHBaseBulkPutExample.java @@ -18,7 +18,6 @@ package org.apache.hadoop.hbase.spark.example.hbasecontext; import java.util.ArrayList; import java.util.List; - import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.TableName; @@ -29,11 +28,13 @@ import org.apache.spark.SparkConf; import org.apache.spark.api.java.JavaRDD; import org.apache.spark.api.java.JavaSparkContext; import org.apache.spark.api.java.function.Function; +import org.apache.yetus.audience.InterfaceAudience; /** * This is a simple example of putting records in HBase * with the bulkPut function. */ +@InterfaceAudience.Private final public class JavaHBaseBulkPutExample { private JavaHBaseBulkPutExample() {} http://git-wip-us.apache.org/repos/asf/hbase/blob/a6eeb26c/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/example/hbasecontext/JavaHBaseDistributedScan.java -- diff --git
[1/4] hbase git commit: HBASE-20212 Make all Public classes have InterfaceAudience category
Repository: hbase Updated Branches: refs/heads/branch-2 0a94c9c25 -> dd9e46bbf http://git-wip-us.apache.org/repos/asf/hbase/blob/dd9e46bb/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/THBaseThreadPoolExecutor.java -- diff --git a/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/THBaseThreadPoolExecutor.java b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/THBaseThreadPoolExecutor.java index 973cad7..c86f476 100644 --- a/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/THBaseThreadPoolExecutor.java +++ b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/THBaseThreadPoolExecutor.java @@ -22,12 +22,14 @@ import java.util.concurrent.BlockingQueue; import java.util.concurrent.ThreadFactory; import java.util.concurrent.ThreadPoolExecutor; import java.util.concurrent.TimeUnit; +import org.apache.yetus.audience.InterfaceAudience; /** * A ThreadPoolExecutor customized for working with HBase thrift to update metrics before and * after the execution of a task. */ +@InterfaceAudience.Private public class THBaseThreadPoolExecutor extends ThreadPoolExecutor { private ThriftMetrics metrics; http://git-wip-us.apache.org/repos/asf/hbase/blob/dd9e46bb/hbase-zookeeper/pom.xml -- diff --git a/hbase-zookeeper/pom.xml b/hbase-zookeeper/pom.xml index 963476a..588afb5 100644 --- a/hbase-zookeeper/pom.xml +++ b/hbase-zookeeper/pom.xml @@ -91,6 +91,10 @@ org.codehaus.mojo findbugs-maven-plugin + +net.revelc.code +warbucks-maven-plugin + http://git-wip-us.apache.org/repos/asf/hbase/blob/dd9e46bb/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/RegionNormalizerTracker.java -- diff --git a/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/RegionNormalizerTracker.java b/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/RegionNormalizerTracker.java index a50ce4c..7413879 100644 --- a/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/RegionNormalizerTracker.java +++ b/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/RegionNormalizerTracker.java @@ -19,10 +19,10 @@ package org.apache.hadoop.hbase.zookeeper; import java.io.IOException; - import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.util.Bytes; +import org.apache.yetus.audience.InterfaceAudience; import org.apache.zookeeper.KeeperException; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -33,6 +33,7 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.RegionNormalizerProtos; /** * Tracks region normalizer state up in ZK */ +@InterfaceAudience.Private public class RegionNormalizerTracker extends ZKNodeTracker { private static final Logger LOG = LoggerFactory.getLogger(RegionNormalizerTracker.class); http://git-wip-us.apache.org/repos/asf/hbase/blob/dd9e46bb/pom.xml -- diff --git a/pom.xml b/pom.xml index 5ca31f8..abbe4ad 100755 --- a/pom.xml +++ b/pom.xml @@ -909,6 +909,34 @@ true + + net.revelc.code + warbucks-maven-plugin + ${maven.warbucks.version} + +false + + + +(?!.*(.generated.|.tmpl.|\$)).* +false +true + false +false +false + org[.]apache[.]yetus[.]audience[.]InterfaceAudience.* + + + + + + run-warbucks + +check + + + + @@ -1382,6 +1410,7 @@ 3.0.0 3.4 3.0.1 +1.1.0 1.5.0.Final 2.8.2 3.2.2
[1/4] hbase git commit: HBASE-20212 Make all Public classes have InterfaceAudience category
Repository: hbase Updated Branches: refs/heads/branch-2.0 9802c1771 -> 70c170727 http://git-wip-us.apache.org/repos/asf/hbase/blob/70c17072/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescer.java -- diff --git a/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescer.java b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescer.java index 3fb7254..f2abe2e 100644 --- a/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescer.java +++ b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescer.java @@ -31,7 +31,6 @@ import java.util.concurrent.ThreadPoolExecutor; import java.util.concurrent.TimeUnit; import java.util.concurrent.atomic.AtomicInteger; import java.util.concurrent.atomic.LongAdder; - import org.apache.hadoop.hbase.CellUtil; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.thrift.ThriftServerRunner.HBaseHandler; @@ -40,6 +39,7 @@ import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Threads; import org.apache.hadoop.metrics2.util.MBeans; import org.apache.thrift.TException; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -52,6 +52,7 @@ import org.slf4j.LoggerFactory; * thrift server dies or is shut down before everything in the queue is drained. * */ +@InterfaceAudience.Private public class IncrementCoalescer implements IncrementCoalescerMBean { /** http://git-wip-us.apache.org/repos/asf/hbase/blob/70c17072/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescerMBean.java -- diff --git a/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescerMBean.java b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescerMBean.java index 604fa97..06cf193 100644 --- a/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescerMBean.java +++ b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/IncrementCoalescerMBean.java @@ -18,6 +18,9 @@ package org.apache.hadoop.hbase.thrift; +import org.apache.yetus.audience.InterfaceAudience; + +@InterfaceAudience.Private public interface IncrementCoalescerMBean { int getQueueSize(); http://git-wip-us.apache.org/repos/asf/hbase/blob/70c17072/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/THBaseThreadPoolExecutor.java -- diff --git a/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/THBaseThreadPoolExecutor.java b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/THBaseThreadPoolExecutor.java index 973cad7..c86f476 100644 --- a/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/THBaseThreadPoolExecutor.java +++ b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/THBaseThreadPoolExecutor.java @@ -22,12 +22,14 @@ import java.util.concurrent.BlockingQueue; import java.util.concurrent.ThreadFactory; import java.util.concurrent.ThreadPoolExecutor; import java.util.concurrent.TimeUnit; +import org.apache.yetus.audience.InterfaceAudience; /** * A ThreadPoolExecutor customized for working with HBase thrift to update metrics before and * after the execution of a task. */ +@InterfaceAudience.Private public class THBaseThreadPoolExecutor extends ThreadPoolExecutor { private ThriftMetrics metrics; http://git-wip-us.apache.org/repos/asf/hbase/blob/70c17072/hbase-zookeeper/pom.xml -- diff --git a/hbase-zookeeper/pom.xml b/hbase-zookeeper/pom.xml index 8c92968..a49dbe7 100644 --- a/hbase-zookeeper/pom.xml +++ b/hbase-zookeeper/pom.xml @@ -91,6 +91,10 @@ org.codehaus.mojo findbugs-maven-plugin + +net.revelc.code +warbucks-maven-plugin + http://git-wip-us.apache.org/repos/asf/hbase/blob/70c17072/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/RegionNormalizerTracker.java -- diff --git a/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/RegionNormalizerTracker.java b/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/RegionNormalizerTracker.java index a50ce4c..7413879 100644 --- a/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/RegionNormalizerTracker.java +++ b/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/RegionNormalizerTracker.java @@ -19,10 +19,10 @@ package org.apache.hadoop.hbase.zookeeper; import java.io.IOException; - import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.exceptions.DeserializationException; import