[03/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.SequentialWriteTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.SequentialWriteTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.SequentialWriteTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.SequentialWriteTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.SequentialWriteTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[32/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AppendTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AppendTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AppendTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AppendTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AppendTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import org.apache.hadoop.hbase.filter.FilterAllFilter;
[13/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange1Test.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange1Test.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange1Test.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange1Test.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange1Test.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import
[50/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/checkstyle-aggregate.html -- diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html index 2e471d7..d807e05 100644 --- a/checkstyle-aggregate.html +++ b/checkstyle-aggregate.html @@ -7,7 +7,7 @@ - + Apache HBase Checkstyle Results @@ -284,7 +284,7 @@ 3609 0 0 -15864 +15865 Files @@ -617,7 +617,7 @@ org/apache/hadoop/hbase/PerformanceEvaluation.java 0 0 -30 +31 org/apache/hadoop/hbase/PerformanceEvaluationCommons.java 0 @@ -10229,7 +10229,7 @@ http://checkstyle.sourceforge.net/config_blocks.html#NeedBraces;>NeedBraces -1936 +1935 Error coding @@ -10322,12 +10322,12 @@ http://checkstyle.sourceforge.net/config_javadoc.html#JavadocTagContinuationIndentation;>JavadocTagContinuationIndentation offset: 2 -783 +797 Error http://checkstyle.sourceforge.net/config_javadoc.html#NonEmptyAtclauseDescription;>NonEmptyAtclauseDescription -3836 +3822 Error misc @@ -10345,7 +10345,7 @@ max: 100 ignorePattern: ^package.*|^import.*|a href|href|http://|https://|ftp://|org.apache.thrift.|com.google.protobuf.|hbase.protobuf.generated -1613 +1615 Error @@ -14536,7 +14536,7 @@ Error javadoc -NonEmptyAtclauseDescription +JavadocTagContinuationIndentation Javadoc comment at column 26 has parse error. Missed HTML close tag 'arg'. Sometimes it means that close tag missed for one of previous tags. 44 @@ -15172,7 +15172,7 @@ Error javadoc -NonEmptyAtclauseDescription +JavadocTagContinuationIndentation Javadoc comment at column 4 has parse error. Missed HTML close tag 'pre'. Sometimes it means that close tag missed for one of previous tags. 59 @@ -16386,373 +16386,373 @@ blocks NeedBraces 'if' construct must use '{}'s. -179 +169 Error blocks NeedBraces 'if' construct must use '{}'s. -287 +277 Error indentation Indentation 'method def rcurly' have incorrect indentation level 1, expected level should be 2. -399 +389 Error javadoc JavadocTagContinuationIndentation Line continuation have incorrect indentation level, expected level should be 2. -455 +445 Error indentation Indentation 'case' child have incorrect indentation level 4, expected level should be 6. -486 +476 Error indentation Indentation 'block' child have incorrect indentation level 6, expected level should be 8. -487 +477 Error indentation Indentation 'block' child have incorrect indentation level 6, expected level should be 8. -488 +478 Error indentation Indentation 'block' child have incorrect indentation level 6, expected level should be 8. -489 +479 Error indentation Indentation 'block' child have incorrect indentation level 6, expected level should be 8. -490 +480 Error indentation Indentation 'block' child have incorrect indentation level 6, expected level should be 8. -491 +481 Error indentation Indentation 'case' child have incorrect indentation level 4, expected level should be 6. -492 +482 Error indentation Indentation 'case' child have incorrect indentation level 4, expected level should be 6. -493 +483 Error indentation Indentation 'case' child have incorrect indentation level 4, expected level should be 6. -494 +484 Error indentation Indentation 'block' child have incorrect indentation level 6, expected level should be 8. -495 +485 Error indentation Indentation 'case' child have incorrect indentation level 4, expected level should be 6. -509 +499 Error indentation Indentation 'block' child have incorrect indentation level 6, expected level should be 8. -510 +500 Error indentation Indentation 'block' child have incorrect indentation level 6, expected level should be 8. -511 +501 Error indentation Indentation 'block' child have incorrect indentation level 6, expected level should be 8. -512 +502 Error indentation Indentation 'block' child have incorrect indentation level 6, expected level should be 8. -513 +503 Error indentation Indentation 'block' child have incorrect indentation level 6, expected level should be 8. -514 +504 Error indentation Indentation 'block' child have incorrect indentation level 6, expected level should be 8. -515 +505 Error indentation Indentation 'case' child have incorrect indentation level 4, expected level should be 6. -516 +506 Error indentation Indentation 'case' child have incorrect indentation level 4, expected level should be 6. -517 +507 Error indentation Indentation 'case' child have incorrect indentation level 4, expected level should be 6. -518 +508 Error indentation Indentation 'block' child have incorrect indentation level 6, expected level should be 8. -519 +509 Error indentation Indentation 'block' child have incorrect indentation level 6, expected level should be 8. -520 +510 Error indentation
[17/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.EvaluationMapTask.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.EvaluationMapTask.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.EvaluationMapTask.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.EvaluationMapTask.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.EvaluationMapTask.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[28/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialReadTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialReadTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialReadTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialReadTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialReadTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter;
[18/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.Counter.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.Counter.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.Counter.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.Counter.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.Counter.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import org.apache.hadoop.hbase.filter.FilterAllFilter; +076import
[01/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
Repository: hbase-site Updated Branches: refs/heads/asf-site 64bc8d859 -> f2065178e http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.TableTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.TableTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.TableTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.TableTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.TableTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import
[11/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange100Test.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange100Test.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange100Test.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange100Test.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange100Test.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import
[35/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.RandomReadTest.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.RandomReadTest.html b/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.RandomReadTest.html index 54e6ab2..ccc5090 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.RandomReadTest.html +++ b/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.RandomReadTest.html @@ -128,7 +128,7 @@ var activeTableTab = "activeTableTab"; -static class PerformanceEvaluation.RandomReadTest +static class PerformanceEvaluation.RandomReadTest extends PerformanceEvaluation.TableTest @@ -247,7 +247,7 @@ extends PerformanceEvaluation.TestBase -generateStatus, getLastRow, getLatencyHistogram, getShortLatencyReport, getShortValueSizeReport, getStartRow, getValueLength, i sRandomValueSize, test, testSetup, testTimed, updateValueSize, updateValueSize, updateValueSize +generateStatus, getLastRow, getLatencyHistogram, getShortLatencyReport, getShortValueSizeReport, getStartRow, getValueLength, i sRandomValueSize, test, testSetup, testTimed, updateScanMetrics, updateValueSize, updateValueSize, updateValueSize @@ -276,7 +276,7 @@ extends consistency -private finalorg.apache.hadoop.hbase.client.Consistency consistency +private finalorg.apache.hadoop.hbase.client.Consistency consistency @@ -285,7 +285,7 @@ extends gets -privatehttps://docs.oracle.com/javase/8/docs/api/java/util/ArrayList.html?is-external=true; title="class or interface in java.util">ArrayListorg.apache.hadoop.hbase.client.Get gets +privatehttps://docs.oracle.com/javase/8/docs/api/java/util/ArrayList.html?is-external=true; title="class or interface in java.util">ArrayListorg.apache.hadoop.hbase.client.Get gets @@ -294,7 +294,7 @@ extends rd -privatehttps://docs.oracle.com/javase/8/docs/api/java/util/Random.html?is-external=true; title="class or interface in java.util">Random rd +privatehttps://docs.oracle.com/javase/8/docs/api/java/util/Random.html?is-external=true; title="class or interface in java.util">Random rd @@ -311,7 +311,7 @@ extends RandomReadTest -RandomReadTest(org.apache.hadoop.hbase.client.Connectioncon, +RandomReadTest(org.apache.hadoop.hbase.client.Connectioncon, PerformanceEvaluation.TestOptionsoptions, PerformanceEvaluation.Statusstatus) @@ -330,7 +330,7 @@ extends testRow -voidtestRow(inti) +voidtestRow(inti) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException, https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true; title="class or interface in java.lang">InterruptedException @@ -348,7 +348,7 @@ extends getReportingPeriod -protectedintgetReportingPeriod() +protectedintgetReportingPeriod() Overrides: getReportingPeriodin classPerformanceEvaluation.TestBase @@ -361,7 +361,7 @@ extends testTakedown -protectedvoidtestTakedown() +protectedvoidtestTakedown() throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Overrides: http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange1Test.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange1Test.html b/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange1Test.html index 8751ac6..c96d41e 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange1Test.html +++ b/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange1Test.html @@ -133,7 +133,7 @@ var activeTableTab = "activeTableTab"; -static class PerformanceEvaluation.RandomScanWithRange1Test +static class PerformanceEvaluation.RandomScanWithRange1Test extends PerformanceEvaluation.RandomScanWithRangeTest @@ -232,7 +232,7 @@ extends PerformanceEvaluation.TestBase -generateStatus, getLastRow, getLatencyHistogram, getShortLatencyReport, getShortValueSizeReport, getStartRow, getValueLength, i sRandomValueSize, test, testSetup, testTakedown, testTimed, updateValueSize, updateValueSize, updateValueSize +generateStatus, getLastRow, getLatencyHistogram, getShortLatencyReport, getShortValueSizeReport, getStartRow, getValueLength, i sRandomValueSize, test, testSetup, testTakedown, testTimed, updateScanMetrics, updateValueSize, updateValueSize, updateValueSize @@ -261,7 +261,7 @@
[26/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTableTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTableTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTableTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTableTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTableTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[31/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncRandomReadTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncRandomReadTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncRandomReadTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncRandomReadTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncRandomReadTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[09/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRangeTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRangeTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRangeTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRangeTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRangeTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter;
[34/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.SequentialWriteTest.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.SequentialWriteTest.html b/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.SequentialWriteTest.html index 28a2264..c2b39f0 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.SequentialWriteTest.html +++ b/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.SequentialWriteTest.html @@ -128,7 +128,7 @@ var activeTableTab = "activeTableTab"; -static class PerformanceEvaluation.SequentialWriteTest +static class PerformanceEvaluation.SequentialWriteTest extends PerformanceEvaluation.BufferedMutatorTest @@ -220,7 +220,7 @@ extends PerformanceEvaluation.TestBase -generateStatus, getLastRow, getLatencyHistogram, getReportingPeriod, getShortLatencyReport, getShortValueSizeReport, getStartRow, getValueLength, isRandomValueSize, test, testSetup, testTakedown, testTimed, updateValueSize, updateValueSize, updateValueSize +generateStatus, getLastRow, getLatencyHistogram, getReportingPeriod, getShortLatencyReport, getShortValueSizeReport, getStartRow, getValueLength, isRandomValueSize, test, testSetup, testTakedown, testTimed, updateScanMetrics, updateValueSize, updateValueSize, updateValueSize @@ -249,7 +249,7 @@ extends SequentialWriteTest -SequentialWriteTest(org.apache.hadoop.hbase.client.Connectioncon, +SequentialWriteTest(org.apache.hadoop.hbase.client.Connectioncon, PerformanceEvaluation.TestOptionsoptions, PerformanceEvaluation.Statusstatus) @@ -268,7 +268,7 @@ extends testRow -voidtestRow(inti) +voidtestRow(inti) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Specified by: http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.TableTest.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.TableTest.html b/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.TableTest.html index 487dc94..5a0a45c 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.TableTest.html +++ b/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.TableTest.html @@ -127,7 +127,7 @@ var activeTableTab = "activeTableTab"; -abstract static class PerformanceEvaluation.TableTest +abstract static class PerformanceEvaluation.TableTest extends PerformanceEvaluation.Test @@ -220,7 +220,7 @@ extends PerformanceEvaluation.TestBase -generateStatus, getLastRow, getLatencyHistogram, getReportingPeriod, getShortLatencyReport, getShortValueSizeReport, getStartRow, getValueLength, isRandomValueSize, test, testRow, testSetup, testTakedown, testTimed, updateValueSize, updateValueSize, updateValueSize +generateStatus, getLastRow, getLatencyHistogram, getReportingPeriod, getShortLatencyReport, getShortValueSizeReport, getStartRow, getValueLength, isRandomValueSize, test, testRow, testSetup, testTakedown, testTimed, updateScanMetrics, updateValueSize, updateValueSize, updateValueSize @@ -249,7 +249,7 @@ extends table -protectedorg.apache.hadoop.hbase.client.Table table +protectedorg.apache.hadoop.hbase.client.Table table @@ -266,7 +266,7 @@ extends TableTest -TableTest(org.apache.hadoop.hbase.client.Connectioncon, +TableTest(org.apache.hadoop.hbase.client.Connectioncon, PerformanceEvaluation.TestOptionsoptions, PerformanceEvaluation.Statusstatus) @@ -285,7 +285,7 @@ extends onStartup -voidonStartup() +voidonStartup() throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Specified by: @@ -301,7 +301,7 @@ extends onTakedown -voidonTakedown() +voidonTakedown() throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Specified by: http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.Test.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.Test.html b/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.Test.html index 47d323f..8d630a4 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.Test.html +++ b/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.Test.html @@ -122,7 +122,7 @@ var activeTableTab = "activeTableTab";
[44/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectingVisitor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectingVisitor.html b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectingVisitor.html index cdd2f36..fea2b5a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectingVisitor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectingVisitor.html @@ -151,2029 +151,2019 @@ 143 private static final Logger LOG = LoggerFactory.getLogger(MetaTableAccessor.class); 144 private static final Logger METALOG = LoggerFactory.getLogger("org.apache.hadoop.hbase.META"); 145 -146 private static final byte[] META_REGION_PREFIX; -147 static { -148// Copy the prefix from FIRST_META_REGIONINFO into META_REGION_PREFIX. -149// FIRST_META_REGIONINFO == 'hbase:meta,,1'. META_REGION_PREFIX == 'hbase:meta,' -150int len = RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName().length - 2; -151META_REGION_PREFIX = new byte [len]; -152 System.arraycopy(RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName(), 0, -153 META_REGION_PREFIX, 0, len); -154 } -155 -156 @VisibleForTesting -157 public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent"); -158 -159 private static final byte ESCAPE_BYTE = (byte) 0xFF; -160 -161 private static final byte SEPARATED_BYTE = 0x00; -162 -163 /** -164 * Lists all of the table regions currently in META. -165 * Deprecated, keep there until some test use this. -166 * @param connection what we will use -167 * @param tableName table to list -168 * @return Map of all user-space regions to servers -169 * @deprecated use {@link #getTableRegionsAndLocations}, region can have multiple locations -170 */ -171 @Deprecated -172 public static NavigableMapRegionInfo, ServerName allTableRegions( -173 Connection connection, final TableName tableName) throws IOException { -174final NavigableMapRegionInfo, ServerName regions = new TreeMap(); -175Visitor visitor = new TableVisitorBase(tableName) { -176 @Override -177 public boolean visitInternal(Result result) throws IOException { -178RegionLocations locations = getRegionLocations(result); -179if (locations == null) return true; -180for (HRegionLocation loc : locations.getRegionLocations()) { -181 if (loc != null) { -182RegionInfo regionInfo = loc.getRegionInfo(); -183regions.put(regionInfo, loc.getServerName()); -184 } -185} -186return true; -187 } -188}; -189scanMetaForTableRegions(connection, visitor, tableName); -190return regions; -191 } -192 -193 @InterfaceAudience.Private -194 public enum QueryType { -195ALL(HConstants.TABLE_FAMILY, HConstants.CATALOG_FAMILY), -196REGION(HConstants.CATALOG_FAMILY), -197TABLE(HConstants.TABLE_FAMILY), -198 REPLICATION(HConstants.REPLICATION_BARRIER_FAMILY); -199 -200private final byte[][] families; -201 -202QueryType(byte[]... families) { -203 this.families = families; -204} -205 -206byte[][] getFamilies() { -207 return this.families; -208} -209 } -210 -211 /** The delimiter for meta columns for replicaIds gt; 0 */ -212 protected static final char META_REPLICA_ID_DELIMITER = '_'; -213 -214 /** A regex for parsing server columns from meta. See above javadoc for meta layout */ -215 private static final Pattern SERVER_COLUMN_PATTERN -216= Pattern.compile("^server(_[0-9a-fA-F]{4})?$"); -217 -218 -219 // Reading operations // -220 -221 -222 /** -223 * Performs a full scan of codehbase:meta/code for regions. -224 * @param connection connection we're using -225 * @param visitor Visitor invoked against each row in regions family. +146 @VisibleForTesting +147 public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent"); +148 +149 private static final byte ESCAPE_BYTE = (byte) 0xFF; +150 +151 private static final byte SEPARATED_BYTE = 0x00; +152 +153 /** +154 * Lists all of the table regions currently in META. +155 * Deprecated, keep there until some test use this. +156 * @param connection what we will use +157 * @param tableName table to list +158 * @return Map of all user-space regions to servers +159 * @deprecated use {@link #getTableRegionsAndLocations}, region can have multiple locations +160 */ +161 @Deprecated +162 public static NavigableMapRegionInfo, ServerName allTableRegions( +163 Connection connection, final TableName tableName) throws IOException { +164final NavigableMapRegionInfo, ServerName regions = new TreeMap(); +165Visitor visitor = new
[16/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.FilteredScanTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.FilteredScanTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.FilteredScanTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.FilteredScanTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.FilteredScanTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[15/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.IncrementTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.IncrementTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.IncrementTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.IncrementTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.IncrementTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[21/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CheckAndMutateTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CheckAndMutateTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CheckAndMutateTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CheckAndMutateTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CheckAndMutateTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[43/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html index cdd2f36..fea2b5a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.DefaultVisitorBase.html @@ -151,2029 +151,2019 @@ 143 private static final Logger LOG = LoggerFactory.getLogger(MetaTableAccessor.class); 144 private static final Logger METALOG = LoggerFactory.getLogger("org.apache.hadoop.hbase.META"); 145 -146 private static final byte[] META_REGION_PREFIX; -147 static { -148// Copy the prefix from FIRST_META_REGIONINFO into META_REGION_PREFIX. -149// FIRST_META_REGIONINFO == 'hbase:meta,,1'. META_REGION_PREFIX == 'hbase:meta,' -150int len = RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName().length - 2; -151META_REGION_PREFIX = new byte [len]; -152 System.arraycopy(RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName(), 0, -153 META_REGION_PREFIX, 0, len); -154 } -155 -156 @VisibleForTesting -157 public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent"); -158 -159 private static final byte ESCAPE_BYTE = (byte) 0xFF; -160 -161 private static final byte SEPARATED_BYTE = 0x00; -162 -163 /** -164 * Lists all of the table regions currently in META. -165 * Deprecated, keep there until some test use this. -166 * @param connection what we will use -167 * @param tableName table to list -168 * @return Map of all user-space regions to servers -169 * @deprecated use {@link #getTableRegionsAndLocations}, region can have multiple locations -170 */ -171 @Deprecated -172 public static NavigableMapRegionInfo, ServerName allTableRegions( -173 Connection connection, final TableName tableName) throws IOException { -174final NavigableMapRegionInfo, ServerName regions = new TreeMap(); -175Visitor visitor = new TableVisitorBase(tableName) { -176 @Override -177 public boolean visitInternal(Result result) throws IOException { -178RegionLocations locations = getRegionLocations(result); -179if (locations == null) return true; -180for (HRegionLocation loc : locations.getRegionLocations()) { -181 if (loc != null) { -182RegionInfo regionInfo = loc.getRegionInfo(); -183regions.put(regionInfo, loc.getServerName()); -184 } -185} -186return true; -187 } -188}; -189scanMetaForTableRegions(connection, visitor, tableName); -190return regions; -191 } -192 -193 @InterfaceAudience.Private -194 public enum QueryType { -195ALL(HConstants.TABLE_FAMILY, HConstants.CATALOG_FAMILY), -196REGION(HConstants.CATALOG_FAMILY), -197TABLE(HConstants.TABLE_FAMILY), -198 REPLICATION(HConstants.REPLICATION_BARRIER_FAMILY); -199 -200private final byte[][] families; -201 -202QueryType(byte[]... families) { -203 this.families = families; -204} -205 -206byte[][] getFamilies() { -207 return this.families; -208} -209 } -210 -211 /** The delimiter for meta columns for replicaIds gt; 0 */ -212 protected static final char META_REPLICA_ID_DELIMITER = '_'; -213 -214 /** A regex for parsing server columns from meta. See above javadoc for meta layout */ -215 private static final Pattern SERVER_COLUMN_PATTERN -216= Pattern.compile("^server(_[0-9a-fA-F]{4})?$"); -217 -218 -219 // Reading operations // -220 -221 -222 /** -223 * Performs a full scan of codehbase:meta/code for regions. -224 * @param connection connection we're using -225 * @param visitor Visitor invoked against each row in regions family. +146 @VisibleForTesting +147 public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent"); +148 +149 private static final byte ESCAPE_BYTE = (byte) 0xFF; +150 +151 private static final byte SEPARATED_BYTE = 0x00; +152 +153 /** +154 * Lists all of the table regions currently in META. +155 * Deprecated, keep there until some test use this. +156 * @param connection what we will use +157 * @param tableName table to list +158 * @return Map of all user-space regions to servers +159 * @deprecated use {@link #getTableRegionsAndLocations}, region can have multiple locations +160 */ +161 @Deprecated +162 public static NavigableMapRegionInfo, ServerName allTableRegions( +163 Connection connection, final TableName tableName) throws IOException { +164final NavigableMapRegionInfo, ServerName regions = new TreeMap(); +165Visitor visitor = new
[51/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07. Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/f2065178 Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/f2065178 Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/f2065178 Branch: refs/heads/asf-site Commit: f2065178e456f32ea5bcb6c5b82571151cb659f5 Parents: 64bc8d8 Author: jenkinsAuthored: Sat May 5 14:51:47 2018 + Committer: jenkins Committed: Sat May 5 14:51:47 2018 + -- acid-semantics.html | 4 +- apache_hbase_reference_guide.pdf| 4 +- book.html | 2 +- bulk-loads.html | 4 +- checkstyle-aggregate.html | 33798 + checkstyle.rss | 4 +- coc.html| 4 +- dependencies.html | 4 +- dependency-convergence.html | 4 +- dependency-info.html| 4 +- dependency-management.html | 4 +- devapidocs/constant-values.html | 6 +- devapidocs/index-all.html | 2 - .../MetaTableAccessor.CloseableVisitor.html | 2 +- .../MetaTableAccessor.CollectAllVisitor.html| 6 +- .../MetaTableAccessor.CollectingVisitor.html|12 +- .../MetaTableAccessor.DefaultVisitorBase.html | 8 +- .../hbase/MetaTableAccessor.QueryType.html |18 +- ...aTableAccessor.ReplicationBarrierResult.html |18 +- .../MetaTableAccessor.TableVisitorBase.html | 8 +- .../hadoop/hbase/MetaTableAccessor.Visitor.html | 4 +- .../apache/hadoop/hbase/MetaTableAccessor.html | 241 +- .../hadoop/hbase/backup/package-tree.html | 4 +- .../hadoop/hbase/client/package-tree.html |26 +- .../hadoop/hbase/filter/package-tree.html |10 +- .../hadoop/hbase/io/hfile/package-tree.html | 4 +- .../apache/hadoop/hbase/ipc/package-tree.html | 4 +- .../hbase/master/balancer/package-tree.html | 2 +- .../hadoop/hbase/master/package-tree.html | 4 +- .../hbase/master/procedure/package-tree.html| 2 +- .../hadoop/hbase/monitoring/package-tree.html | 2 +- .../org/apache/hadoop/hbase/package-tree.html |14 +- .../hadoop/hbase/procedure2/package-tree.html | 4 +- .../hadoop/hbase/quotas/package-tree.html | 6 +- .../hadoop/hbase/regionserver/package-tree.html |16 +- .../regionserver/querymatcher/package-tree.html | 2 +- .../hbase/regionserver/wal/package-tree.html| 2 +- .../replication/regionserver/package-tree.html | 2 +- .../hbase/security/access/package-tree.html | 2 +- .../hadoop/hbase/security/package-tree.html | 2 +- .../hadoop/hbase/thrift/package-tree.html | 2 +- .../apache/hadoop/hbase/util/package-tree.html | 8 +- .../MetaTableAccessor.CloseableVisitor.html | 3970 +- .../MetaTableAccessor.CollectAllVisitor.html| 3970 +- .../MetaTableAccessor.CollectingVisitor.html| 3970 +- .../MetaTableAccessor.DefaultVisitorBase.html | 3970 +- .../hbase/MetaTableAccessor.QueryType.html | 3970 +- ...aTableAccessor.ReplicationBarrierResult.html | 3970 +- .../MetaTableAccessor.TableVisitorBase.html | 3970 +- .../hadoop/hbase/MetaTableAccessor.Visitor.html | 3970 +- .../apache/hadoop/hbase/MetaTableAccessor.html | 3970 +- .../org/apache/hadoop/hbase/Version.html| 6 +- downloads.html | 4 +- export_control.html | 4 +- index.html | 4 +- integration.html| 4 +- issue-tracking.html | 4 +- license.html| 4 +- mail-lists.html | 4 +- metrics.html| 4 +- old_news.html | 4 +- plugin-management.html | 4 +- plugins.html| 4 +- poweredbyhbase.html | 4 +- project-info.html | 4 +- project-reports.html| 4 +- project-summary.html| 4 +- pseudo-distributed.html | 4 +- replication.html| 4 +- resources.html | 4 +- source-repository.html | 4 +- sponsors.html | 4 +-
[30/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncRandomWriteTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncRandomWriteTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncRandomWriteTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncRandomWriteTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncRandomWriteTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[04/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.SequentialReadTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.SequentialReadTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.SequentialReadTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.SequentialReadTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.SequentialReadTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[47/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html b/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html index 5890b1a..9b88fec 100644 --- a/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html @@ -167,10 +167,10 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.backup.BackupRestoreConstants.BackupCommand org.apache.hadoop.hbase.backup.BackupInfo.BackupPhase -org.apache.hadoop.hbase.backup.BackupInfo.BackupState org.apache.hadoop.hbase.backup.BackupType +org.apache.hadoop.hbase.backup.BackupRestoreConstants.BackupCommand +org.apache.hadoop.hbase.backup.BackupInfo.BackupState http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/devapidocs/org/apache/hadoop/hbase/client/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/package-tree.html b/devapidocs/org/apache/hadoop/hbase/client/package-tree.html index ed476e8..e17c4b2 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/client/package-tree.html @@ -550,24 +550,24 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.client.RequestController.ReturnCode -org.apache.hadoop.hbase.client.Durability -org.apache.hadoop.hbase.client.Consistency -org.apache.hadoop.hbase.client.CompactionState -org.apache.hadoop.hbase.client.ScannerCallable.MoreResults -org.apache.hadoop.hbase.client.AsyncProcessTask.SubmittedRows org.apache.hadoop.hbase.client.AsyncRequestFutureImpl.Retry -org.apache.hadoop.hbase.client.MobCompactPartitionPolicy -org.apache.hadoop.hbase.client.AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState -org.apache.hadoop.hbase.client.AsyncScanSingleRegionRpcRetryingCaller.ScanControllerState -org.apache.hadoop.hbase.client.CompactType -org.apache.hadoop.hbase.client.IsolationLevel +org.apache.hadoop.hbase.client.AsyncProcessTask.SubmittedRows org.apache.hadoop.hbase.client.MasterSwitchType -org.apache.hadoop.hbase.client.RegionLocateType org.apache.hadoop.hbase.client.AbstractResponse.ResponseType -org.apache.hadoop.hbase.client.SnapshotType +org.apache.hadoop.hbase.client.CompactType +org.apache.hadoop.hbase.client.Consistency +org.apache.hadoop.hbase.client.AsyncScanSingleRegionRpcRetryingCaller.ScanControllerState +org.apache.hadoop.hbase.client.ScannerCallable.MoreResults org.apache.hadoop.hbase.client.TableState.State org.apache.hadoop.hbase.client.Scan.ReadType +org.apache.hadoop.hbase.client.AsyncScanSingleRegionRpcRetryingCaller.ScanResumerState +org.apache.hadoop.hbase.client.RegionLocateType +org.apache.hadoop.hbase.client.MobCompactPartitionPolicy +org.apache.hadoop.hbase.client.CompactionState +org.apache.hadoop.hbase.client.IsolationLevel +org.apache.hadoop.hbase.client.SnapshotType +org.apache.hadoop.hbase.client.RequestController.ReturnCode +org.apache.hadoop.hbase.client.Durability http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/devapidocs/org/apache/hadoop/hbase/filter/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/filter/package-tree.html b/devapidocs/org/apache/hadoop/hbase/filter/package-tree.html index bf0f6e6..62c0444 100644 --- a/devapidocs/org/apache/hadoop/hbase/filter/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/filter/package-tree.html @@ -183,14 +183,14 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in
[45/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html index cdd2f36..fea2b5a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CollectAllVisitor.html @@ -151,2029 +151,2019 @@ 143 private static final Logger LOG = LoggerFactory.getLogger(MetaTableAccessor.class); 144 private static final Logger METALOG = LoggerFactory.getLogger("org.apache.hadoop.hbase.META"); 145 -146 private static final byte[] META_REGION_PREFIX; -147 static { -148// Copy the prefix from FIRST_META_REGIONINFO into META_REGION_PREFIX. -149// FIRST_META_REGIONINFO == 'hbase:meta,,1'. META_REGION_PREFIX == 'hbase:meta,' -150int len = RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName().length - 2; -151META_REGION_PREFIX = new byte [len]; -152 System.arraycopy(RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName(), 0, -153 META_REGION_PREFIX, 0, len); -154 } -155 -156 @VisibleForTesting -157 public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent"); -158 -159 private static final byte ESCAPE_BYTE = (byte) 0xFF; -160 -161 private static final byte SEPARATED_BYTE = 0x00; -162 -163 /** -164 * Lists all of the table regions currently in META. -165 * Deprecated, keep there until some test use this. -166 * @param connection what we will use -167 * @param tableName table to list -168 * @return Map of all user-space regions to servers -169 * @deprecated use {@link #getTableRegionsAndLocations}, region can have multiple locations -170 */ -171 @Deprecated -172 public static NavigableMapRegionInfo, ServerName allTableRegions( -173 Connection connection, final TableName tableName) throws IOException { -174final NavigableMapRegionInfo, ServerName regions = new TreeMap(); -175Visitor visitor = new TableVisitorBase(tableName) { -176 @Override -177 public boolean visitInternal(Result result) throws IOException { -178RegionLocations locations = getRegionLocations(result); -179if (locations == null) return true; -180for (HRegionLocation loc : locations.getRegionLocations()) { -181 if (loc != null) { -182RegionInfo regionInfo = loc.getRegionInfo(); -183regions.put(regionInfo, loc.getServerName()); -184 } -185} -186return true; -187 } -188}; -189scanMetaForTableRegions(connection, visitor, tableName); -190return regions; -191 } -192 -193 @InterfaceAudience.Private -194 public enum QueryType { -195ALL(HConstants.TABLE_FAMILY, HConstants.CATALOG_FAMILY), -196REGION(HConstants.CATALOG_FAMILY), -197TABLE(HConstants.TABLE_FAMILY), -198 REPLICATION(HConstants.REPLICATION_BARRIER_FAMILY); -199 -200private final byte[][] families; -201 -202QueryType(byte[]... families) { -203 this.families = families; -204} -205 -206byte[][] getFamilies() { -207 return this.families; -208} -209 } -210 -211 /** The delimiter for meta columns for replicaIds gt; 0 */ -212 protected static final char META_REPLICA_ID_DELIMITER = '_'; -213 -214 /** A regex for parsing server columns from meta. See above javadoc for meta layout */ -215 private static final Pattern SERVER_COLUMN_PATTERN -216= Pattern.compile("^server(_[0-9a-fA-F]{4})?$"); -217 -218 -219 // Reading operations // -220 -221 -222 /** -223 * Performs a full scan of codehbase:meta/code for regions. -224 * @param connection connection we're using -225 * @param visitor Visitor invoked against each row in regions family. +146 @VisibleForTesting +147 public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent"); +148 +149 private static final byte ESCAPE_BYTE = (byte) 0xFF; +150 +151 private static final byte SEPARATED_BYTE = 0x00; +152 +153 /** +154 * Lists all of the table regions currently in META. +155 * Deprecated, keep there until some test use this. +156 * @param connection what we will use +157 * @param tableName table to list +158 * @return Map of all user-space regions to servers +159 * @deprecated use {@link #getTableRegionsAndLocations}, region can have multiple locations +160 */ +161 @Deprecated +162 public static NavigableMapRegionInfo, ServerName allTableRegions( +163 Connection connection, final TableName tableName) throws IOException { +164final NavigableMapRegionInfo, ServerName regions = new TreeMap(); +165Visitor visitor = new
[49/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/checkstyle.rss -- diff --git a/checkstyle.rss b/checkstyle.rss index e6f25dd..62c7425 100644 --- a/checkstyle.rss +++ b/checkstyle.rss @@ -26,7 +26,7 @@ under the License. 2007 - 2018 The Apache Software Foundation File: 3609, - Errors: 15864, + Errors: 15865, Warnings: 0, Infos: 0 @@ -4605,7 +4605,7 @@ under the License. 0 - 30 + 31 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/coc.html -- diff --git a/coc.html b/coc.html index d4ae2af..c8ad6c7 100644 --- a/coc.html +++ b/coc.html @@ -7,7 +7,7 @@ - + Apache HBase Code of Conduct Policy @@ -375,7 +375,7 @@ email to mailto:priv...@hbase.apache.org;>the priv https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-05-04 + Last Published: 2018-05-05 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/dependencies.html -- diff --git a/dependencies.html b/dependencies.html index 78ca6ef..bf02801 100644 --- a/dependencies.html +++ b/dependencies.html @@ -7,7 +7,7 @@ - + Apache HBase Project Dependencies @@ -440,7 +440,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-05-04 + Last Published: 2018-05-05 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/dependency-convergence.html -- diff --git a/dependency-convergence.html b/dependency-convergence.html index 05ba8a3..5f2f21f 100644 --- a/dependency-convergence.html +++ b/dependency-convergence.html @@ -7,7 +7,7 @@ - + Apache HBase Reactor Dependency Convergence @@ -1105,7 +1105,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-05-04 + Last Published: 2018-05-05 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/dependency-info.html -- diff --git a/dependency-info.html b/dependency-info.html index 19f0fec..703c94f 100644 --- a/dependency-info.html +++ b/dependency-info.html @@ -7,7 +7,7 @@ - + Apache HBase Dependency Information @@ -313,7 +313,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-05-04 + Last Published: 2018-05-05 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/dependency-management.html -- diff --git a/dependency-management.html b/dependency-management.html index ad19247..59ef175 100644 --- a/dependency-management.html +++ b/dependency-management.html @@ -7,7 +7,7 @@ - + Apache HBase Project Dependency Management @@ -969,7 +969,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-05-04 + Last Published: 2018-05-05 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/devapidocs/constant-values.html -- diff --git a/devapidocs/constant-values.html b/devapidocs/constant-values.html index 5f72784..b81f234 100644 --- a/devapidocs/constant-values.html +++ b/devapidocs/constant-values.html @@ -3768,21 +3768,21 @@ publicstaticfinalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String date -"Fri May 4 14:39:10 UTC 2018" +"Sat May 5 14:43:16 UTC 2018" publicstaticfinalhttps://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">String revision
[36/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html b/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html index e2b381d..379ef34 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html +++ b/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html @@ -128,7 +128,7 @@ var activeTableTab = "activeTableTab"; -static class PerformanceEvaluation.AsyncSequentialWriteTest +static class PerformanceEvaluation.AsyncSequentialWriteTest extends PerformanceEvaluation.AsyncTableTest @@ -220,7 +220,7 @@ extends PerformanceEvaluation.TestBase -generateStatus, getLastRow, getLatencyHistogram, getReportingPeriod, getShortLatencyReport, getShortValueSizeReport, getStartRow, getValueLength, isRandomValueSize, test, testSetup, testTakedown, testTimed, updateValueSize, updateValueSize, updateValueSize +generateStatus, getLastRow, getLatencyHistogram, getReportingPeriod, getShortLatencyReport, getShortValueSizeReport, getStartRow, getValueLength, isRandomValueSize, test, testSetup, testTakedown, testTimed, updateScanMetrics, updateValueSize, updateValueSize, updateValueSize @@ -249,7 +249,7 @@ extends AsyncSequentialWriteTest -AsyncSequentialWriteTest(org.apache.hadoop.hbase.client.AsyncConnectioncon, +AsyncSequentialWriteTest(org.apache.hadoop.hbase.client.AsyncConnectioncon, PerformanceEvaluation.TestOptionsoptions, PerformanceEvaluation.Statusstatus) @@ -268,7 +268,7 @@ extends testRow -voidtestRow(inti) +voidtestRow(inti) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException, https://docs.oracle.com/javase/8/docs/api/java/lang/InterruptedException.html?is-external=true; title="class or interface in java.lang">InterruptedException http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTableTest.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTableTest.html b/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTableTest.html index cfcf2af..271bda9 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTableTest.html +++ b/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTableTest.html @@ -127,7 +127,7 @@ var activeTableTab = "activeTableTab"; -abstract static class PerformanceEvaluation.AsyncTableTest +abstract static class PerformanceEvaluation.AsyncTableTest extends PerformanceEvaluation.AsyncTest @@ -220,7 +220,7 @@ extends PerformanceEvaluation.TestBase -generateStatus, getLastRow, getLatencyHistogram, getReportingPeriod, getShortLatencyReport, getShortValueSizeReport, getStartRow, getValueLength, isRandomValueSize, test, testRow, testSetup, testTakedown, testTimed, updateValueSize, updateValueSize, updateValueSize +generateStatus, getLastRow, getLatencyHistogram, getReportingPeriod, getShortLatencyReport, getShortValueSizeReport, getStartRow, getValueLength, isRandomValueSize, test, testRow, testSetup, testTakedown, testTimed, updateScanMetrics, updateValueSize, updateValueSize, updateValueSize @@ -249,7 +249,7 @@ extends table -protectedorg.apache.hadoop.hbase.client.AsyncTable? table +protectedorg.apache.hadoop.hbase.client.AsyncTable? table @@ -266,7 +266,7 @@ extends AsyncTableTest -AsyncTableTest(org.apache.hadoop.hbase.client.AsyncConnectioncon, +AsyncTableTest(org.apache.hadoop.hbase.client.AsyncConnectioncon, PerformanceEvaluation.TestOptionsoptions, PerformanceEvaluation.Statusstatus) @@ -285,7 +285,7 @@ extends onStartup -voidonStartup() +voidonStartup() throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Specified by: @@ -301,7 +301,7 @@ extends onTakedown -voidonTakedown() +voidonTakedown() throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Specified by: http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTest.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTest.html
[07/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomWriteTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomWriteTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomWriteTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomWriteTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomWriteTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[23/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CASTableTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CASTableTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CASTableTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CASTableTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CASTableTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[37/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/devapidocs/src-html/org/apache/hadoop/hbase/Version.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/Version.html b/devapidocs/src-html/org/apache/hadoop/hbase/Version.html index 0423a39..534acbb 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/Version.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/Version.html @@ -16,11 +16,11 @@ 008@InterfaceAudience.Private 009public class Version { 010 public static final String version = "3.0.0-SNAPSHOT"; -011 public static final String revision = "87f5b5f3411d96c31b4cb61b9a57ced22be91d1f"; +011 public static final String revision = "acd0d1e446c164d9c54bfb461b2d449c8d717c07"; 012 public static final String user = "jenkins"; -013 public static final String date = "Fri May 4 14:39:10 UTC 2018"; +013 public static final String date = "Sat May 5 14:43:16 UTC 2018"; 014 public static final String url = "git://jenkins-websites1.apache.org/home/jenkins/jenkins-slave/workspace/hbase_generate_website/hbase"; -015 public static final String srcChecksum = "9cfe3eb8d49902aafbb6d8066c40fb45"; +015 public static final String srcChecksum = "ad489a437d1ef798b4c0329dd68b451a"; 016} http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/downloads.html -- diff --git a/downloads.html b/downloads.html index 7fad404..9a2b856 100644 --- a/downloads.html +++ b/downloads.html @@ -7,7 +7,7 @@ - + Apache HBase Apache HBase Downloads @@ -366,7 +366,7 @@ under the License. --> https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-05-04 + Last Published: 2018-05-05 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/export_control.html -- diff --git a/export_control.html b/export_control.html index e0a5a62..4e37cae 100644 --- a/export_control.html +++ b/export_control.html @@ -7,7 +7,7 @@ - + Apache HBase Export Control @@ -331,7 +331,7 @@ for more details. https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-05-04 + Last Published: 2018-05-05 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/index.html -- diff --git a/index.html b/index.html index 9fcef97..64420ab 100644 --- a/index.html +++ b/index.html @@ -7,7 +7,7 @@ - + Apache HBase Apache HBase™ Home @@ -409,7 +409,7 @@ Apache HBase is an open-source, distributed, versioned, non-relational database https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-05-04 + Last Published: 2018-05-05 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/integration.html -- diff --git a/integration.html b/integration.html index f53166f..1576246 100644 --- a/integration.html +++ b/integration.html @@ -7,7 +7,7 @@ - + Apache HBase CI Management @@ -291,7 +291,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-05-04 + Last Published: 2018-05-05 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/issue-tracking.html -- diff --git a/issue-tracking.html b/issue-tracking.html index cf58b27..56a8126 100644 --- a/issue-tracking.html +++ b/issue-tracking.html @@ -7,7 +7,7 @@ - + Apache HBase Issue Management @@ -288,7 +288,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2018-05-04 + Last Published: 2018-05-05 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/license.html -- diff --git a/license.html b/license.html index 435b1e2..fb236a4 100644 --- a/license.html +++
[22/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CheckAndDeleteTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CheckAndDeleteTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CheckAndDeleteTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CheckAndDeleteTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CheckAndDeleteTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[12/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange1000Test.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange1000Test.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange1000Test.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange1000Test.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange1000Test.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import
[20/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CheckAndPutTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CheckAndPutTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CheckAndPutTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CheckAndPutTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CheckAndPutTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[08/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomSeekScanTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomSeekScanTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomSeekScanTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomSeekScanTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomSeekScanTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
hbase git commit: HBASE-20531 RS may throw NPE when close meta regions in shutdown procedure.
Repository: hbase Updated Branches: refs/heads/master acd0d1e44 -> 971f5350e HBASE-20531 RS may throw NPE when close meta regions in shutdown procedure. Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/971f5350 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/971f5350 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/971f5350 Branch: refs/heads/master Commit: 971f5350e81591e9360677f3617a399f453b6b96 Parents: acd0d1e Author: huzhengAuthored: Fri May 4 19:44:10 2018 +0800 Committer: huzheng Committed: Sun May 6 11:18:23 2018 +0800 -- .../hbase/regionserver/HRegionServer.java | 22 ++-- 1 file changed, 11 insertions(+), 11 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/971f5350/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java index 240de85..188aef6 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java @@ -1052,15 +1052,6 @@ public class HRegionServer extends HasThread implements if (this.storefileRefresher != null) this.storefileRefresher.cancel(true); sendShutdownInterrupt(); -// Stop the quota manager -if (rsQuotaManager != null) { - rsQuotaManager.stop(); -} -if (rsSpaceQuotaManager != null) { - rsSpaceQuotaManager.stop(); - rsSpaceQuotaManager = null; -} - // Stop the snapshot and other procedure handlers, forcefully killing all running tasks if (rspmHost != null) { rspmHost.stop(this.abortRequested || this.killed); @@ -1106,6 +1097,15 @@ public class HRegionServer extends HasThread implements LOG.info("stopping server " + this.serverName + "; all regions closed."); } +// Stop the quota manager +if (rsQuotaManager != null) { + rsQuotaManager.stop(); +} +if (rsSpaceQuotaManager != null) { + rsSpaceQuotaManager.stop(); + rsSpaceQuotaManager = null; +} + //fsOk flag may be changed when closing regions throws exception. if (this.fsOk) { shutdownWAL(!abortRequested); @@ -3697,9 +3697,9 @@ public class HRegionServer extends HasThread implements @Override public boolean reportFileArchivalForQuotas(TableName tableName, - Collection > archivedFiles) { + Collection > archivedFiles) { RegionServerStatusService.BlockingInterface rss = rssStub; -if (rss == null) { +if (rss == null || rsSpaceQuotaManager == null) { // the current server could be stopping. LOG.trace("Skipping file archival reporting to HMaster as stub is null"); return false;
hbase git commit: HBASE-20527 Remove unused code in MetaTableAccessor
Repository: hbase Updated Branches: refs/heads/master 291dedbf8 -> acd0d1e44 HBASE-20527 Remove unused code in MetaTableAccessor Signed-off-by: Chia-Ping TsaiProject: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/acd0d1e4 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/acd0d1e4 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/acd0d1e4 Branch: refs/heads/master Commit: acd0d1e446c164d9c54bfb461b2d449c8d717c07 Parents: 291dedb Author: Mingdao Yang Authored: Sat May 5 01:58:02 2018 + Committer: Chia-Ping Tsai Committed: Sat May 5 22:15:54 2018 +0800 -- .../java/org/apache/hadoop/hbase/MetaTableAccessor.java | 10 -- 1 file changed, 10 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/acd0d1e4/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java index e8ce811..91f3cf7 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java @@ -143,16 +143,6 @@ public class MetaTableAccessor { private static final Logger LOG = LoggerFactory.getLogger(MetaTableAccessor.class); private static final Logger METALOG = LoggerFactory.getLogger("org.apache.hadoop.hbase.META"); - private static final byte[] META_REGION_PREFIX; - static { -// Copy the prefix from FIRST_META_REGIONINFO into META_REGION_PREFIX. -// FIRST_META_REGIONINFO == 'hbase:meta,,1'. META_REGION_PREFIX == 'hbase:meta,' -int len = RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName().length - 2; -META_REGION_PREFIX = new byte [len]; -System.arraycopy(RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName(), 0, - META_REGION_PREFIX, 0, len); - } - @VisibleForTesting public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent");
hbase git commit: HBASE-20527 Remove unused code in MetaTableAccessor
Repository: hbase Updated Branches: refs/heads/branch-2.0 81f69e585 -> d77989e9f HBASE-20527 Remove unused code in MetaTableAccessor Signed-off-by: Chia-Ping TsaiProject: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/d77989e9 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/d77989e9 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/d77989e9 Branch: refs/heads/branch-2.0 Commit: d77989e9f7b51f058d09694e5109e3aec05e0596 Parents: 81f69e5 Author: Mingdao Yang Authored: Sat May 5 22:26:12 2018 +0800 Committer: Chia-Ping Tsai Committed: Sat May 5 22:27:27 2018 +0800 -- .../java/org/apache/hadoop/hbase/MetaTableAccessor.java | 10 -- 1 file changed, 10 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/d77989e9/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java index d6bbf53..1880a0d 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java @@ -137,16 +137,6 @@ public class MetaTableAccessor { private static final Logger LOG = LoggerFactory.getLogger(MetaTableAccessor.class); private static final Logger METALOG = LoggerFactory.getLogger("org.apache.hadoop.hbase.META"); - static final byte [] META_REGION_PREFIX; - static { -// Copy the prefix from FIRST_META_REGIONINFO into META_REGION_PREFIX. -// FIRST_META_REGIONINFO == 'hbase:meta,,1'. META_REGION_PREFIX == 'hbase:meta,' -int len = RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName().length - 2; -META_REGION_PREFIX = new byte [len]; -System.arraycopy(RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName(), 0, - META_REGION_PREFIX, 0, len); - } - /** * Lists all of the table regions currently in META. * Deprecated, keep there until some test use this.
hbase git commit: HBASE-20527 Remove unused code in MetaTableAccessor
Repository: hbase Updated Branches: refs/heads/branch-2 991d78ca2 -> 2373451f9 HBASE-20527 Remove unused code in MetaTableAccessor Signed-off-by: Chia-Ping TsaiProject: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/2373451f Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/2373451f Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/2373451f Branch: refs/heads/branch-2 Commit: 2373451f94d8a71483213a82368031123a9d4815 Parents: 991d78c Author: Mingdao Yang Authored: Sat May 5 01:58:02 2018 + Committer: Chia-Ping Tsai Committed: Sat May 5 22:20:24 2018 +0800 -- .../java/org/apache/hadoop/hbase/MetaTableAccessor.java | 10 -- 1 file changed, 10 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/2373451f/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java index e8ce811..91f3cf7 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java @@ -143,16 +143,6 @@ public class MetaTableAccessor { private static final Logger LOG = LoggerFactory.getLogger(MetaTableAccessor.class); private static final Logger METALOG = LoggerFactory.getLogger("org.apache.hadoop.hbase.META"); - private static final byte[] META_REGION_PREFIX; - static { -// Copy the prefix from FIRST_META_REGIONINFO into META_REGION_PREFIX. -// FIRST_META_REGIONINFO == 'hbase:meta,,1'. META_REGION_PREFIX == 'hbase:meta,' -int len = RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName().length - 2; -META_REGION_PREFIX = new byte [len]; -System.arraycopy(RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName(), 0, - META_REGION_PREFIX, 0, len); - } - @VisibleForTesting public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent");
[29/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncScanTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncScanTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncScanTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncScanTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncScanTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[19/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CmdDescriptor.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CmdDescriptor.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CmdDescriptor.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CmdDescriptor.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.CmdDescriptor.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[38/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.html b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.html index cdd2f36..fea2b5a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.html @@ -151,2029 +151,2019 @@ 143 private static final Logger LOG = LoggerFactory.getLogger(MetaTableAccessor.class); 144 private static final Logger METALOG = LoggerFactory.getLogger("org.apache.hadoop.hbase.META"); 145 -146 private static final byte[] META_REGION_PREFIX; -147 static { -148// Copy the prefix from FIRST_META_REGIONINFO into META_REGION_PREFIX. -149// FIRST_META_REGIONINFO == 'hbase:meta,,1'. META_REGION_PREFIX == 'hbase:meta,' -150int len = RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName().length - 2; -151META_REGION_PREFIX = new byte [len]; -152 System.arraycopy(RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName(), 0, -153 META_REGION_PREFIX, 0, len); -154 } -155 -156 @VisibleForTesting -157 public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent"); -158 -159 private static final byte ESCAPE_BYTE = (byte) 0xFF; -160 -161 private static final byte SEPARATED_BYTE = 0x00; -162 -163 /** -164 * Lists all of the table regions currently in META. -165 * Deprecated, keep there until some test use this. -166 * @param connection what we will use -167 * @param tableName table to list -168 * @return Map of all user-space regions to servers -169 * @deprecated use {@link #getTableRegionsAndLocations}, region can have multiple locations -170 */ -171 @Deprecated -172 public static NavigableMapRegionInfo, ServerName allTableRegions( -173 Connection connection, final TableName tableName) throws IOException { -174final NavigableMapRegionInfo, ServerName regions = new TreeMap(); -175Visitor visitor = new TableVisitorBase(tableName) { -176 @Override -177 public boolean visitInternal(Result result) throws IOException { -178RegionLocations locations = getRegionLocations(result); -179if (locations == null) return true; -180for (HRegionLocation loc : locations.getRegionLocations()) { -181 if (loc != null) { -182RegionInfo regionInfo = loc.getRegionInfo(); -183regions.put(regionInfo, loc.getServerName()); -184 } -185} -186return true; -187 } -188}; -189scanMetaForTableRegions(connection, visitor, tableName); -190return regions; -191 } -192 -193 @InterfaceAudience.Private -194 public enum QueryType { -195ALL(HConstants.TABLE_FAMILY, HConstants.CATALOG_FAMILY), -196REGION(HConstants.CATALOG_FAMILY), -197TABLE(HConstants.TABLE_FAMILY), -198 REPLICATION(HConstants.REPLICATION_BARRIER_FAMILY); -199 -200private final byte[][] families; -201 -202QueryType(byte[]... families) { -203 this.families = families; -204} -205 -206byte[][] getFamilies() { -207 return this.families; -208} -209 } -210 -211 /** The delimiter for meta columns for replicaIds gt; 0 */ -212 protected static final char META_REPLICA_ID_DELIMITER = '_'; -213 -214 /** A regex for parsing server columns from meta. See above javadoc for meta layout */ -215 private static final Pattern SERVER_COLUMN_PATTERN -216= Pattern.compile("^server(_[0-9a-fA-F]{4})?$"); -217 -218 -219 // Reading operations // -220 -221 -222 /** -223 * Performs a full scan of codehbase:meta/code for regions. -224 * @param connection connection we're using -225 * @param visitor Visitor invoked against each row in regions family. +146 @VisibleForTesting +147 public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent"); +148 +149 private static final byte ESCAPE_BYTE = (byte) 0xFF; +150 +151 private static final byte SEPARATED_BYTE = 0x00; +152 +153 /** +154 * Lists all of the table regions currently in META. +155 * Deprecated, keep there until some test use this. +156 * @param connection what we will use +157 * @param tableName table to list +158 * @return Map of all user-space regions to servers +159 * @deprecated use {@link #getTableRegionsAndLocations}, region can have multiple locations +160 */ +161 @Deprecated +162 public static NavigableMapRegionInfo, ServerName allTableRegions( +163 Connection connection, final TableName tableName) throws IOException { +164final NavigableMapRegionInfo, ServerName regions = new TreeMap(); +165Visitor visitor = new TableVisitorBase(tableName) { +166 @Override +167 public boolean visitInternal(Result result)
hbase-site git commit: INFRA-10751 Empty commit
Repository: hbase-site Updated Branches: refs/heads/asf-site f2065178e -> 6f3be57d5 INFRA-10751 Empty commit Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/6f3be57d Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/6f3be57d Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/6f3be57d Branch: refs/heads/asf-site Commit: 6f3be57d5e3349f419b8fe009542cb9d4bfb6b2c Parents: f206517 Author: jenkinsAuthored: Sat May 5 14:52:12 2018 + Committer: jenkins Committed: Sat May 5 14:52:12 2018 + -- --
[27/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncSequentialWriteTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import
[48/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/devapidocs/org/apache/hadoop/hbase/MetaTableAccessor.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/MetaTableAccessor.html b/devapidocs/org/apache/hadoop/hbase/MetaTableAccessor.html index 4ebc5d4..9090907 100644 --- a/devapidocs/org/apache/hadoop/hbase/MetaTableAccessor.html +++ b/devapidocs/org/apache/hadoop/hbase/MetaTableAccessor.html @@ -253,28 +253,24 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html LOG -private static byte[] -META_REGION_PREFIX - - protected static char META_REPLICA_ID_DELIMITER The delimiter for meta columns for replicaIds 0 - + private static org.slf4j.Logger METALOG - + static byte[] REPLICATION_PARENT_QUALIFIER - + private static byte SEPARATED_BYTE - + private static https://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Pattern SERVER_COLUMN_PATTERN A regex for parsing server columns from meta. @@ -1065,22 +1061,13 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html private static finalorg.slf4j.Logger METALOG - - - - - -META_REGION_PREFIX -private static finalbyte[] META_REGION_PREFIX - - REPLICATION_PARENT_QUALIFIER -public static finalbyte[] REPLICATION_PARENT_QUALIFIER +public static finalbyte[] REPLICATION_PARENT_QUALIFIER @@ -1089,7 +1076,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html ESCAPE_BYTE -private static finalbyte ESCAPE_BYTE +private static finalbyte ESCAPE_BYTE See Also: Constant Field Values @@ -1102,7 +1089,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html SEPARATED_BYTE -private static finalbyte SEPARATED_BYTE +private static finalbyte SEPARATED_BYTE See Also: Constant Field Values @@ -1115,7 +1102,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html META_REPLICA_ID_DELIMITER -protected static finalchar META_REPLICA_ID_DELIMITER +protected static finalchar META_REPLICA_ID_DELIMITER The delimiter for meta columns for replicaIds 0 See Also: @@ -1129,7 +1116,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html SERVER_COLUMN_PATTERN -private static finalhttps://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Pattern SERVER_COLUMN_PATTERN +private static finalhttps://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Pattern SERVER_COLUMN_PATTERN A regex for parsing server columns from meta. See above javadoc for meta layout @@ -1165,7 +1152,7 @@ extends https://docs.oracle.com/javase/8/docs/api/java/lang/Object.html allTableRegions https://docs.oracle.com/javase/8/docs/api/java/lang/Deprecated.html?is-external=true; title="class or interface in java.lang">@Deprecated -public statichttps://docs.oracle.com/javase/8/docs/api/java/util/NavigableMap.html?is-external=true; title="class or interface in java.util">NavigableMapRegionInfo,ServerNameallTableRegions(Connectionconnection, +public statichttps://docs.oracle.com/javase/8/docs/api/java/util/NavigableMap.html?is-external=true; title="class or interface in java.util">NavigableMapRegionInfo,ServerNameallTableRegions(Connectionconnection, TableNametableName) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Deprecated.use getTableRegionsAndLocations(org.apache.hadoop.hbase.client.Connection, org.apache.hadoop.hbase.TableName), region can have multiple locations @@ -1188,7 +1175,7 @@ public statichttps://docs.oracle.com/javase/8/docs/api/java/util/ fullScanRegions -public staticvoidfullScanRegions(Connectionconnection, +public staticvoidfullScanRegions(Connectionconnection, MetaTableAccessor.Visitorvisitor) throws https://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Performs a full scan of hbase:meta for regions. @@ -1207,7 +1194,7 @@ public statichttps://docs.oracle.com/javase/8/docs/api/java/util/ fullScanRegions -public statichttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListResultfullScanRegions(Connectionconnection) +public statichttps://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListResultfullScanRegions(Connectionconnection)
[05/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.ScanTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.ScanTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.ScanTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.ScanTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.ScanTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import org.apache.hadoop.hbase.filter.FilterAllFilter; +076import
[14/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomReadTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomReadTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomReadTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomReadTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomReadTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[25/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.AsyncTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import org.apache.hadoop.hbase.filter.FilterAllFilter; +076import
[41/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.ReplicationBarrierResult.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.ReplicationBarrierResult.html b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.ReplicationBarrierResult.html index cdd2f36..fea2b5a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.ReplicationBarrierResult.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.ReplicationBarrierResult.html @@ -151,2029 +151,2019 @@ 143 private static final Logger LOG = LoggerFactory.getLogger(MetaTableAccessor.class); 144 private static final Logger METALOG = LoggerFactory.getLogger("org.apache.hadoop.hbase.META"); 145 -146 private static final byte[] META_REGION_PREFIX; -147 static { -148// Copy the prefix from FIRST_META_REGIONINFO into META_REGION_PREFIX. -149// FIRST_META_REGIONINFO == 'hbase:meta,,1'. META_REGION_PREFIX == 'hbase:meta,' -150int len = RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName().length - 2; -151META_REGION_PREFIX = new byte [len]; -152 System.arraycopy(RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName(), 0, -153 META_REGION_PREFIX, 0, len); -154 } -155 -156 @VisibleForTesting -157 public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent"); -158 -159 private static final byte ESCAPE_BYTE = (byte) 0xFF; -160 -161 private static final byte SEPARATED_BYTE = 0x00; -162 -163 /** -164 * Lists all of the table regions currently in META. -165 * Deprecated, keep there until some test use this. -166 * @param connection what we will use -167 * @param tableName table to list -168 * @return Map of all user-space regions to servers -169 * @deprecated use {@link #getTableRegionsAndLocations}, region can have multiple locations -170 */ -171 @Deprecated -172 public static NavigableMapRegionInfo, ServerName allTableRegions( -173 Connection connection, final TableName tableName) throws IOException { -174final NavigableMapRegionInfo, ServerName regions = new TreeMap(); -175Visitor visitor = new TableVisitorBase(tableName) { -176 @Override -177 public boolean visitInternal(Result result) throws IOException { -178RegionLocations locations = getRegionLocations(result); -179if (locations == null) return true; -180for (HRegionLocation loc : locations.getRegionLocations()) { -181 if (loc != null) { -182RegionInfo regionInfo = loc.getRegionInfo(); -183regions.put(regionInfo, loc.getServerName()); -184 } -185} -186return true; -187 } -188}; -189scanMetaForTableRegions(connection, visitor, tableName); -190return regions; -191 } -192 -193 @InterfaceAudience.Private -194 public enum QueryType { -195ALL(HConstants.TABLE_FAMILY, HConstants.CATALOG_FAMILY), -196REGION(HConstants.CATALOG_FAMILY), -197TABLE(HConstants.TABLE_FAMILY), -198 REPLICATION(HConstants.REPLICATION_BARRIER_FAMILY); -199 -200private final byte[][] families; -201 -202QueryType(byte[]... families) { -203 this.families = families; -204} -205 -206byte[][] getFamilies() { -207 return this.families; -208} -209 } -210 -211 /** The delimiter for meta columns for replicaIds gt; 0 */ -212 protected static final char META_REPLICA_ID_DELIMITER = '_'; -213 -214 /** A regex for parsing server columns from meta. See above javadoc for meta layout */ -215 private static final Pattern SERVER_COLUMN_PATTERN -216= Pattern.compile("^server(_[0-9a-fA-F]{4})?$"); -217 -218 -219 // Reading operations // -220 -221 -222 /** -223 * Performs a full scan of codehbase:meta/code for regions. -224 * @param connection connection we're using -225 * @param visitor Visitor invoked against each row in regions family. +146 @VisibleForTesting +147 public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent"); +148 +149 private static final byte ESCAPE_BYTE = (byte) 0xFF; +150 +151 private static final byte SEPARATED_BYTE = 0x00; +152 +153 /** +154 * Lists all of the table regions currently in META. +155 * Deprecated, keep there until some test use this. +156 * @param connection what we will use +157 * @param tableName table to list +158 * @return Map of all user-space regions to servers +159 * @deprecated use {@link #getTableRegionsAndLocations}, region can have multiple locations +160 */ +161 @Deprecated +162 public static NavigableMapRegionInfo, ServerName allTableRegions( +163 Connection connection, final TableName tableName) throws IOException { +164final NavigableMapRegionInfo, ServerName regions = new TreeMap(); +165
[06/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RunResult.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RunResult.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RunResult.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RunResult.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RunResult.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import org.apache.hadoop.hbase.filter.FilterAllFilter; +076import
[40/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.TableVisitorBase.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.TableVisitorBase.html b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.TableVisitorBase.html index cdd2f36..fea2b5a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.TableVisitorBase.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.TableVisitorBase.html @@ -151,2029 +151,2019 @@ 143 private static final Logger LOG = LoggerFactory.getLogger(MetaTableAccessor.class); 144 private static final Logger METALOG = LoggerFactory.getLogger("org.apache.hadoop.hbase.META"); 145 -146 private static final byte[] META_REGION_PREFIX; -147 static { -148// Copy the prefix from FIRST_META_REGIONINFO into META_REGION_PREFIX. -149// FIRST_META_REGIONINFO == 'hbase:meta,,1'. META_REGION_PREFIX == 'hbase:meta,' -150int len = RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName().length - 2; -151META_REGION_PREFIX = new byte [len]; -152 System.arraycopy(RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName(), 0, -153 META_REGION_PREFIX, 0, len); -154 } -155 -156 @VisibleForTesting -157 public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent"); -158 -159 private static final byte ESCAPE_BYTE = (byte) 0xFF; -160 -161 private static final byte SEPARATED_BYTE = 0x00; -162 -163 /** -164 * Lists all of the table regions currently in META. -165 * Deprecated, keep there until some test use this. -166 * @param connection what we will use -167 * @param tableName table to list -168 * @return Map of all user-space regions to servers -169 * @deprecated use {@link #getTableRegionsAndLocations}, region can have multiple locations -170 */ -171 @Deprecated -172 public static NavigableMapRegionInfo, ServerName allTableRegions( -173 Connection connection, final TableName tableName) throws IOException { -174final NavigableMapRegionInfo, ServerName regions = new TreeMap(); -175Visitor visitor = new TableVisitorBase(tableName) { -176 @Override -177 public boolean visitInternal(Result result) throws IOException { -178RegionLocations locations = getRegionLocations(result); -179if (locations == null) return true; -180for (HRegionLocation loc : locations.getRegionLocations()) { -181 if (loc != null) { -182RegionInfo regionInfo = loc.getRegionInfo(); -183regions.put(regionInfo, loc.getServerName()); -184 } -185} -186return true; -187 } -188}; -189scanMetaForTableRegions(connection, visitor, tableName); -190return regions; -191 } -192 -193 @InterfaceAudience.Private -194 public enum QueryType { -195ALL(HConstants.TABLE_FAMILY, HConstants.CATALOG_FAMILY), -196REGION(HConstants.CATALOG_FAMILY), -197TABLE(HConstants.TABLE_FAMILY), -198 REPLICATION(HConstants.REPLICATION_BARRIER_FAMILY); -199 -200private final byte[][] families; -201 -202QueryType(byte[]... families) { -203 this.families = families; -204} -205 -206byte[][] getFamilies() { -207 return this.families; -208} -209 } -210 -211 /** The delimiter for meta columns for replicaIds gt; 0 */ -212 protected static final char META_REPLICA_ID_DELIMITER = '_'; -213 -214 /** A regex for parsing server columns from meta. See above javadoc for meta layout */ -215 private static final Pattern SERVER_COLUMN_PATTERN -216= Pattern.compile("^server(_[0-9a-fA-F]{4})?$"); -217 -218 -219 // Reading operations // -220 -221 -222 /** -223 * Performs a full scan of codehbase:meta/code for regions. -224 * @param connection connection we're using -225 * @param visitor Visitor invoked against each row in regions family. +146 @VisibleForTesting +147 public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent"); +148 +149 private static final byte ESCAPE_BYTE = (byte) 0xFF; +150 +151 private static final byte SEPARATED_BYTE = 0x00; +152 +153 /** +154 * Lists all of the table regions currently in META. +155 * Deprecated, keep there until some test use this. +156 * @param connection what we will use +157 * @param tableName table to list +158 * @return Map of all user-space regions to servers +159 * @deprecated use {@link #getTableRegionsAndLocations}, region can have multiple locations +160 */ +161 @Deprecated +162 public static NavigableMapRegionInfo, ServerName allTableRegions( +163 Connection connection, final TableName tableName) throws IOException { +164final NavigableMapRegionInfo, ServerName regions = new TreeMap(); +165Visitor visitor = new
[39/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.Visitor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.Visitor.html b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.Visitor.html index cdd2f36..fea2b5a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.Visitor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.Visitor.html @@ -151,2029 +151,2019 @@ 143 private static final Logger LOG = LoggerFactory.getLogger(MetaTableAccessor.class); 144 private static final Logger METALOG = LoggerFactory.getLogger("org.apache.hadoop.hbase.META"); 145 -146 private static final byte[] META_REGION_PREFIX; -147 static { -148// Copy the prefix from FIRST_META_REGIONINFO into META_REGION_PREFIX. -149// FIRST_META_REGIONINFO == 'hbase:meta,,1'. META_REGION_PREFIX == 'hbase:meta,' -150int len = RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName().length - 2; -151META_REGION_PREFIX = new byte [len]; -152 System.arraycopy(RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName(), 0, -153 META_REGION_PREFIX, 0, len); -154 } -155 -156 @VisibleForTesting -157 public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent"); -158 -159 private static final byte ESCAPE_BYTE = (byte) 0xFF; -160 -161 private static final byte SEPARATED_BYTE = 0x00; -162 -163 /** -164 * Lists all of the table regions currently in META. -165 * Deprecated, keep there until some test use this. -166 * @param connection what we will use -167 * @param tableName table to list -168 * @return Map of all user-space regions to servers -169 * @deprecated use {@link #getTableRegionsAndLocations}, region can have multiple locations -170 */ -171 @Deprecated -172 public static NavigableMapRegionInfo, ServerName allTableRegions( -173 Connection connection, final TableName tableName) throws IOException { -174final NavigableMapRegionInfo, ServerName regions = new TreeMap(); -175Visitor visitor = new TableVisitorBase(tableName) { -176 @Override -177 public boolean visitInternal(Result result) throws IOException { -178RegionLocations locations = getRegionLocations(result); -179if (locations == null) return true; -180for (HRegionLocation loc : locations.getRegionLocations()) { -181 if (loc != null) { -182RegionInfo regionInfo = loc.getRegionInfo(); -183regions.put(regionInfo, loc.getServerName()); -184 } -185} -186return true; -187 } -188}; -189scanMetaForTableRegions(connection, visitor, tableName); -190return regions; -191 } -192 -193 @InterfaceAudience.Private -194 public enum QueryType { -195ALL(HConstants.TABLE_FAMILY, HConstants.CATALOG_FAMILY), -196REGION(HConstants.CATALOG_FAMILY), -197TABLE(HConstants.TABLE_FAMILY), -198 REPLICATION(HConstants.REPLICATION_BARRIER_FAMILY); -199 -200private final byte[][] families; -201 -202QueryType(byte[]... families) { -203 this.families = families; -204} -205 -206byte[][] getFamilies() { -207 return this.families; -208} -209 } -210 -211 /** The delimiter for meta columns for replicaIds gt; 0 */ -212 protected static final char META_REPLICA_ID_DELIMITER = '_'; -213 -214 /** A regex for parsing server columns from meta. See above javadoc for meta layout */ -215 private static final Pattern SERVER_COLUMN_PATTERN -216= Pattern.compile("^server(_[0-9a-fA-F]{4})?$"); -217 -218 -219 // Reading operations // -220 -221 -222 /** -223 * Performs a full scan of codehbase:meta/code for regions. -224 * @param connection connection we're using -225 * @param visitor Visitor invoked against each row in regions family. +146 @VisibleForTesting +147 public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent"); +148 +149 private static final byte ESCAPE_BYTE = (byte) 0xFF; +150 +151 private static final byte SEPARATED_BYTE = 0x00; +152 +153 /** +154 * Lists all of the table regions currently in META. +155 * Deprecated, keep there until some test use this. +156 * @param connection what we will use +157 * @param tableName table to list +158 * @return Map of all user-space regions to servers +159 * @deprecated use {@link #getTableRegionsAndLocations}, region can have multiple locations +160 */ +161 @Deprecated +162 public static NavigableMapRegionInfo, ServerName allTableRegions( +163 Connection connection, final TableName tableName) throws IOException { +164final NavigableMapRegionInfo, ServerName regions = new TreeMap(); +165Visitor visitor = new TableVisitorBase(tableName) { +166 @Override +167
[24/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.BufferedMutatorTest.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.BufferedMutatorTest.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.BufferedMutatorTest.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.BufferedMutatorTest.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.BufferedMutatorTest.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import
[33/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html b/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html index 87ba170..6be1f69 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html +++ b/testdevapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html @@ -655,11 +655,11 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.regionserver.TestMultiLogThreshold.ActionType -org.apache.hadoop.hbase.regionserver.TestRegionServerReadRequestMetrics.Metric org.apache.hadoop.hbase.regionserver.TestAtomicOperation.TestStep -org.apache.hadoop.hbase.regionserver.TestCacheOnWriteInSchema.CacheOnWriteType +org.apache.hadoop.hbase.regionserver.TestRegionServerReadRequestMetrics.Metric +org.apache.hadoop.hbase.regionserver.TestMultiLogThreshold.ActionType org.apache.hadoop.hbase.regionserver.DataBlockEncodingTool.Manipulation +org.apache.hadoop.hbase.regionserver.TestCacheOnWriteInSchema.CacheOnWriteType http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/org/apache/hadoop/hbase/test/package-tree.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/test/package-tree.html b/testdevapidocs/org/apache/hadoop/hbase/test/package-tree.html index ecfaabf..1a43271 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/test/package-tree.html +++ b/testdevapidocs/org/apache/hadoop/hbase/test/package-tree.html @@ -254,9 +254,9 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) org.apache.hadoop.hbase.test.IntegrationTestLoadAndVerify.Counters -org.apache.hadoop.hbase.test.IntegrationTestWithCellVisibilityLoadAndVerify.Counters -org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList.Verify.Counts org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList.Generator.Counts +org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList.Verify.Counts +org.apache.hadoop.hbase.test.IntegrationTestWithCellVisibilityLoadAndVerify.Counters http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/org/apache/hadoop/hbase/wal/package-tree.html -- diff --git a/testdevapidocs/org/apache/hadoop/hbase/wal/package-tree.html b/testdevapidocs/org/apache/hadoop/hbase/wal/package-tree.html index f7d6df6..4edd3f8 100644 --- a/testdevapidocs/org/apache/hadoop/hbase/wal/package-tree.html +++ b/testdevapidocs/org/apache/hadoop/hbase/wal/package-tree.html @@ -140,8 +140,8 @@ java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.https://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.https://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) org.apache.hadoop.hbase.wal.IOTestProvider.AllowedOperations -org.apache.hadoop.hbase.wal.TestWALSplit.Corruptions org.apache.hadoop.hbase.wal.FaultyFSLog.FailureType +org.apache.hadoop.hbase.wal.TestWALSplit.Corruptions
[02/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.Status.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.Status.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.Status.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.Status.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.Status.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import org.apache.hadoop.hbase.filter.Filter; +075import org.apache.hadoop.hbase.filter.FilterAllFilter; +076import
[10/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange10Test.html -- diff --git a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange10Test.html b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange10Test.html index 2510283..418c60c 100644 --- a/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange10Test.html +++ b/testdevapidocs/src-html/org/apache/hadoop/hbase/PerformanceEvaluation.RandomScanWithRange10Test.html @@ -77,77 +77,77 @@ 069import org.apache.hadoop.hbase.client.RowMutations; 070import org.apache.hadoop.hbase.client.Scan; 071import org.apache.hadoop.hbase.client.Table; -072import org.apache.hadoop.hbase.filter.BinaryComparator; -073import org.apache.hadoop.hbase.filter.Filter; -074import org.apache.hadoop.hbase.filter.FilterAllFilter; -075import org.apache.hadoop.hbase.filter.FilterList; -076import org.apache.hadoop.hbase.filter.PageFilter; -077import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; -078import org.apache.hadoop.hbase.filter.WhileMatchFilter; -079import org.apache.hadoop.hbase.io.compress.Compression; -080import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; -081import org.apache.hadoop.hbase.io.hfile.RandomDistribution; -082import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; -083import org.apache.hadoop.hbase.regionserver.BloomType; -084import org.apache.hadoop.hbase.regionserver.CompactingMemStore; -085import org.apache.hadoop.hbase.trace.HBaseHTraceConfiguration; -086import org.apache.hadoop.hbase.trace.SpanReceiverHost; -087import org.apache.hadoop.hbase.trace.TraceUtil; -088import org.apache.hadoop.hbase.util.ByteArrayHashKey; -089import org.apache.hadoop.hbase.util.Bytes; -090import org.apache.hadoop.hbase.util.Hash; -091import org.apache.hadoop.hbase.util.MurmurHash; -092import org.apache.hadoop.hbase.util.Pair; -093import org.apache.hadoop.hbase.util.YammerHistogramUtils; -094import org.apache.hadoop.io.LongWritable; -095import org.apache.hadoop.io.Text; -096import org.apache.hadoop.mapreduce.Job; -097import org.apache.hadoop.mapreduce.Mapper; -098import org.apache.hadoop.mapreduce.lib.input.NLineInputFormat; -099import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat; -100import org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer; -101import org.apache.hadoop.util.Tool; -102import org.apache.hadoop.util.ToolRunner; -103import org.apache.htrace.core.ProbabilitySampler; -104import org.apache.htrace.core.Sampler; -105import org.apache.htrace.core.TraceScope; -106import org.apache.yetus.audience.InterfaceAudience; -107import org.slf4j.Logger; -108import org.slf4j.LoggerFactory; -109import org.apache.hbase.thirdparty.com.google.common.base.MoreObjects; -110import org.apache.hbase.thirdparty.com.google.common.util.concurrent.ThreadFactoryBuilder; -111 -112/** -113 * Script used evaluating HBase performance and scalability. Runs a HBase -114 * client that steps through one of a set of hardcoded tests or 'experiments' -115 * (e.g. a random reads test, a random writes test, etc.). Pass on the -116 * command-line which test to run and how many clients are participating in -117 * this experiment. Run {@code PerformanceEvaluation --help} to obtain usage. -118 * -119 * pThis class sets up and runs the evaluation programs described in -120 * Section 7, iPerformance Evaluation/i, of the a -121 * href="http://labs.google.com/papers/bigtable.html"Bigtable/a; -122 * paper, pages 8-10. -123 * -124 * pBy default, runs as a mapreduce job where each mapper runs a single test -125 * client. Can also run as a non-mapreduce, multithreaded application by -126 * specifying {@code --nomapred}. Each client does about 1GB of data, unless -127 * specified otherwise. -128 */ -129@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.TOOLS) -130public class PerformanceEvaluation extends Configured implements Tool { -131 static final String RANDOM_SEEK_SCAN = "randomSeekScan"; -132 static final String RANDOM_READ = "randomRead"; -133 private static final Logger LOG = LoggerFactory.getLogger(PerformanceEvaluation.class.getName()); -134 private static final ObjectMapper MAPPER = new ObjectMapper(); -135 static { -136 MAPPER.configure(MapperFeature.SORT_PROPERTIES_ALPHABETICALLY, true); -137 } -138 -139 public static final String TABLE_NAME = "TestTable"; -140 public static final byte[] FAMILY_NAME = Bytes.toBytes("info"); -141 public static final byte [] COLUMN_ZERO = Bytes.toBytes("" + 0); -142 public static final byte [] QUALIFIER_NAME = COLUMN_ZERO; +072import org.apache.hadoop.hbase.client.metrics.ScanMetrics; +073import org.apache.hadoop.hbase.filter.BinaryComparator; +074import
[46/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CloseableVisitor.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CloseableVisitor.html b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CloseableVisitor.html index cdd2f36..fea2b5a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CloseableVisitor.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.CloseableVisitor.html @@ -151,2029 +151,2019 @@ 143 private static final Logger LOG = LoggerFactory.getLogger(MetaTableAccessor.class); 144 private static final Logger METALOG = LoggerFactory.getLogger("org.apache.hadoop.hbase.META"); 145 -146 private static final byte[] META_REGION_PREFIX; -147 static { -148// Copy the prefix from FIRST_META_REGIONINFO into META_REGION_PREFIX. -149// FIRST_META_REGIONINFO == 'hbase:meta,,1'. META_REGION_PREFIX == 'hbase:meta,' -150int len = RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName().length - 2; -151META_REGION_PREFIX = new byte [len]; -152 System.arraycopy(RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName(), 0, -153 META_REGION_PREFIX, 0, len); -154 } -155 -156 @VisibleForTesting -157 public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent"); -158 -159 private static final byte ESCAPE_BYTE = (byte) 0xFF; -160 -161 private static final byte SEPARATED_BYTE = 0x00; -162 -163 /** -164 * Lists all of the table regions currently in META. -165 * Deprecated, keep there until some test use this. -166 * @param connection what we will use -167 * @param tableName table to list -168 * @return Map of all user-space regions to servers -169 * @deprecated use {@link #getTableRegionsAndLocations}, region can have multiple locations -170 */ -171 @Deprecated -172 public static NavigableMapRegionInfo, ServerName allTableRegions( -173 Connection connection, final TableName tableName) throws IOException { -174final NavigableMapRegionInfo, ServerName regions = new TreeMap(); -175Visitor visitor = new TableVisitorBase(tableName) { -176 @Override -177 public boolean visitInternal(Result result) throws IOException { -178RegionLocations locations = getRegionLocations(result); -179if (locations == null) return true; -180for (HRegionLocation loc : locations.getRegionLocations()) { -181 if (loc != null) { -182RegionInfo regionInfo = loc.getRegionInfo(); -183regions.put(regionInfo, loc.getServerName()); -184 } -185} -186return true; -187 } -188}; -189scanMetaForTableRegions(connection, visitor, tableName); -190return regions; -191 } -192 -193 @InterfaceAudience.Private -194 public enum QueryType { -195ALL(HConstants.TABLE_FAMILY, HConstants.CATALOG_FAMILY), -196REGION(HConstants.CATALOG_FAMILY), -197TABLE(HConstants.TABLE_FAMILY), -198 REPLICATION(HConstants.REPLICATION_BARRIER_FAMILY); -199 -200private final byte[][] families; -201 -202QueryType(byte[]... families) { -203 this.families = families; -204} -205 -206byte[][] getFamilies() { -207 return this.families; -208} -209 } -210 -211 /** The delimiter for meta columns for replicaIds gt; 0 */ -212 protected static final char META_REPLICA_ID_DELIMITER = '_'; -213 -214 /** A regex for parsing server columns from meta. See above javadoc for meta layout */ -215 private static final Pattern SERVER_COLUMN_PATTERN -216= Pattern.compile("^server(_[0-9a-fA-F]{4})?$"); -217 -218 -219 // Reading operations // -220 -221 -222 /** -223 * Performs a full scan of codehbase:meta/code for regions. -224 * @param connection connection we're using -225 * @param visitor Visitor invoked against each row in regions family. +146 @VisibleForTesting +147 public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent"); +148 +149 private static final byte ESCAPE_BYTE = (byte) 0xFF; +150 +151 private static final byte SEPARATED_BYTE = 0x00; +152 +153 /** +154 * Lists all of the table regions currently in META. +155 * Deprecated, keep there until some test use this. +156 * @param connection what we will use +157 * @param tableName table to list +158 * @return Map of all user-space regions to servers +159 * @deprecated use {@link #getTableRegionsAndLocations}, region can have multiple locations +160 */ +161 @Deprecated +162 public static NavigableMapRegionInfo, ServerName allTableRegions( +163 Connection connection, final TableName tableName) throws IOException { +164final NavigableMapRegionInfo, ServerName regions = new TreeMap(); +165Visitor visitor = new
[42/51] [partial] hbase-site git commit: Published site at acd0d1e446c164d9c54bfb461b2d449c8d717c07.
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/f2065178/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.QueryType.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.QueryType.html b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.QueryType.html index cdd2f36..fea2b5a 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.QueryType.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/MetaTableAccessor.QueryType.html @@ -151,2029 +151,2019 @@ 143 private static final Logger LOG = LoggerFactory.getLogger(MetaTableAccessor.class); 144 private static final Logger METALOG = LoggerFactory.getLogger("org.apache.hadoop.hbase.META"); 145 -146 private static final byte[] META_REGION_PREFIX; -147 static { -148// Copy the prefix from FIRST_META_REGIONINFO into META_REGION_PREFIX. -149// FIRST_META_REGIONINFO == 'hbase:meta,,1'. META_REGION_PREFIX == 'hbase:meta,' -150int len = RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName().length - 2; -151META_REGION_PREFIX = new byte [len]; -152 System.arraycopy(RegionInfoBuilder.FIRST_META_REGIONINFO.getRegionName(), 0, -153 META_REGION_PREFIX, 0, len); -154 } -155 -156 @VisibleForTesting -157 public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent"); -158 -159 private static final byte ESCAPE_BYTE = (byte) 0xFF; -160 -161 private static final byte SEPARATED_BYTE = 0x00; -162 -163 /** -164 * Lists all of the table regions currently in META. -165 * Deprecated, keep there until some test use this. -166 * @param connection what we will use -167 * @param tableName table to list -168 * @return Map of all user-space regions to servers -169 * @deprecated use {@link #getTableRegionsAndLocations}, region can have multiple locations -170 */ -171 @Deprecated -172 public static NavigableMapRegionInfo, ServerName allTableRegions( -173 Connection connection, final TableName tableName) throws IOException { -174final NavigableMapRegionInfo, ServerName regions = new TreeMap(); -175Visitor visitor = new TableVisitorBase(tableName) { -176 @Override -177 public boolean visitInternal(Result result) throws IOException { -178RegionLocations locations = getRegionLocations(result); -179if (locations == null) return true; -180for (HRegionLocation loc : locations.getRegionLocations()) { -181 if (loc != null) { -182RegionInfo regionInfo = loc.getRegionInfo(); -183regions.put(regionInfo, loc.getServerName()); -184 } -185} -186return true; -187 } -188}; -189scanMetaForTableRegions(connection, visitor, tableName); -190return regions; -191 } -192 -193 @InterfaceAudience.Private -194 public enum QueryType { -195ALL(HConstants.TABLE_FAMILY, HConstants.CATALOG_FAMILY), -196REGION(HConstants.CATALOG_FAMILY), -197TABLE(HConstants.TABLE_FAMILY), -198 REPLICATION(HConstants.REPLICATION_BARRIER_FAMILY); -199 -200private final byte[][] families; -201 -202QueryType(byte[]... families) { -203 this.families = families; -204} -205 -206byte[][] getFamilies() { -207 return this.families; -208} -209 } -210 -211 /** The delimiter for meta columns for replicaIds gt; 0 */ -212 protected static final char META_REPLICA_ID_DELIMITER = '_'; -213 -214 /** A regex for parsing server columns from meta. See above javadoc for meta layout */ -215 private static final Pattern SERVER_COLUMN_PATTERN -216= Pattern.compile("^server(_[0-9a-fA-F]{4})?$"); -217 -218 -219 // Reading operations // -220 -221 -222 /** -223 * Performs a full scan of codehbase:meta/code for regions. -224 * @param connection connection we're using -225 * @param visitor Visitor invoked against each row in regions family. +146 @VisibleForTesting +147 public static final byte[] REPLICATION_PARENT_QUALIFIER = Bytes.toBytes("parent"); +148 +149 private static final byte ESCAPE_BYTE = (byte) 0xFF; +150 +151 private static final byte SEPARATED_BYTE = 0x00; +152 +153 /** +154 * Lists all of the table regions currently in META. +155 * Deprecated, keep there until some test use this. +156 * @param connection what we will use +157 * @param tableName table to list +158 * @return Map of all user-space regions to servers +159 * @deprecated use {@link #getTableRegionsAndLocations}, region can have multiple locations +160 */ +161 @Deprecated +162 public static NavigableMapRegionInfo, ServerName allTableRegions( +163 Connection connection, final TableName tableName) throws IOException { +164final NavigableMapRegionInfo, ServerName regions = new TreeMap(); +165Visitor visitor = new TableVisitorBase(tableName) { +166 @Override