hbase git commit: HBASE-19017 [AMv2] EnableTableProcedure is not retaining the assignments (Ram)
Repository: hbase Updated Branches: refs/heads/branch-2 cf550c2be -> 8b8f7a017 HBASE-19017 [AMv2] EnableTableProcedure is not retaining the assignments (Ram) Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/8b8f7a01 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/8b8f7a01 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/8b8f7a01 Branch: refs/heads/branch-2 Commit: 8b8f7a017b05f055a926546e60eab1a238d567cc Parents: cf550c2 Author: RamkrishnaAuthored: Tue Oct 17 11:06:51 2017 +0530 Committer: Ramkrishna Committed: Tue Oct 17 11:08:41 2017 +0530 -- .../master/assignment/AssignProcedure.java | 4 + .../TestRegionReplicasAreDistributed.java | 186 +++ 2 files changed, 190 insertions(+) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/8b8f7a01/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignProcedure.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignProcedure.java index 01a5f83..3eba571 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignProcedure.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignProcedure.java @@ -198,6 +198,10 @@ public class AssignProcedure extends RegionTransitionProcedure { // Try and keep the location we had before we offlined. retain = true; regionNode.setRegionLocation(lastRegionLocation); +} else if (regionNode.getLastHost() != null) { + retain = true; + LOG.info("Setting lastHost as the region location " + regionNode.getLastHost()); + regionNode.setRegionLocation(regionNode.getLastHost()); } } } http://git-wip-us.apache.org/repos/asf/hbase/blob/8b8f7a01/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicasAreDistributed.java -- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicasAreDistributed.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicasAreDistributed.java new file mode 100644 index 000..d6b7e6f --- /dev/null +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicasAreDistributed.java @@ -0,0 +1,186 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.hadoop.hbase.regionserver; + +import static org.junit.Assert.assertTrue; + +import java.io.IOException; +import java.util.ArrayList; +import java.util.Collection; +import java.util.HashMap; +import java.util.Map; + +import org.apache.commons.logging.Log; +import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.hbase.HBaseTestingUtility; +import org.apache.hadoop.hbase.HConstants; +import org.apache.hadoop.hbase.HTableDescriptor; +import org.apache.hadoop.hbase.ServerName; +import org.apache.hadoop.hbase.TableName; +import org.apache.hadoop.hbase.client.RegionInfo; +import org.apache.hadoop.hbase.client.Table; +import org.apache.hadoop.hbase.testclassification.MediumTests; +import org.apache.hadoop.hbase.testclassification.RegionServerTests; +import org.apache.hadoop.hbase.util.Bytes; +import org.apache.hadoop.hbase.util.RegionSplitter; +import org.junit.AfterClass; +import org.junit.BeforeClass; +import org.junit.Test; +import org.junit.experimental.categories.Category; + +@Category({ RegionServerTests.class, MediumTests.class }) +public class TestRegionReplicasAreDistributed { + + private static final Log LOG = LogFactory.getLog(TestRegionReplicasAreDistributed.class); + + private static final int NB_SERVERS = 3; + private static Table table; + +
hbase git commit: HBASE-19017 [AMv2] EnableTableProcedure is not retaining the assignments (Ram)
Repository: hbase Updated Branches: refs/heads/master be4570fc8 -> 9a27ac878 HBASE-19017 [AMv2] EnableTableProcedure is not retaining the assignments (Ram) Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/9a27ac87 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/9a27ac87 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/9a27ac87 Branch: refs/heads/master Commit: 9a27ac8781ea09ad0890c3f0a6e6676a785818cf Parents: be4570f Author: RamkrishnaAuthored: Tue Oct 17 11:06:51 2017 +0530 Committer: Ramkrishna Committed: Tue Oct 17 11:06:51 2017 +0530 -- .../master/assignment/AssignProcedure.java | 4 + .../TestRegionReplicasAreDistributed.java | 186 +++ 2 files changed, 190 insertions(+) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/9a27ac87/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignProcedure.java -- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignProcedure.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignProcedure.java index 01a5f83..3eba571 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignProcedure.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignProcedure.java @@ -198,6 +198,10 @@ public class AssignProcedure extends RegionTransitionProcedure { // Try and keep the location we had before we offlined. retain = true; regionNode.setRegionLocation(lastRegionLocation); +} else if (regionNode.getLastHost() != null) { + retain = true; + LOG.info("Setting lastHost as the region location " + regionNode.getLastHost()); + regionNode.setRegionLocation(regionNode.getLastHost()); } } } http://git-wip-us.apache.org/repos/asf/hbase/blob/9a27ac87/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicasAreDistributed.java -- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicasAreDistributed.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicasAreDistributed.java new file mode 100644 index 000..d6b7e6f --- /dev/null +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicasAreDistributed.java @@ -0,0 +1,186 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.hadoop.hbase.regionserver; + +import static org.junit.Assert.assertTrue; + +import java.io.IOException; +import java.util.ArrayList; +import java.util.Collection; +import java.util.HashMap; +import java.util.Map; + +import org.apache.commons.logging.Log; +import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.hbase.HBaseTestingUtility; +import org.apache.hadoop.hbase.HConstants; +import org.apache.hadoop.hbase.HTableDescriptor; +import org.apache.hadoop.hbase.ServerName; +import org.apache.hadoop.hbase.TableName; +import org.apache.hadoop.hbase.client.RegionInfo; +import org.apache.hadoop.hbase.client.Table; +import org.apache.hadoop.hbase.testclassification.MediumTests; +import org.apache.hadoop.hbase.testclassification.RegionServerTests; +import org.apache.hadoop.hbase.util.Bytes; +import org.apache.hadoop.hbase.util.RegionSplitter; +import org.junit.AfterClass; +import org.junit.BeforeClass; +import org.junit.Test; +import org.junit.experimental.categories.Category; + +@Category({ RegionServerTests.class, MediumTests.class }) +public class TestRegionReplicasAreDistributed { + + private static final Log LOG = LogFactory.getLog(TestRegionReplicasAreDistributed.class); + + private static final int NB_SERVERS = 3; + private static Table table; + + private
hbase git commit: HBASE-19016 Coordinate storage policy property name for table schema and bulkload
Repository: hbase Updated Branches: refs/heads/branch-2 58b0585d6 -> cf550c2be HBASE-19016 Coordinate storage policy property name for table schema and bulkload Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/cf550c2b Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/cf550c2b Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/cf550c2b Branch: refs/heads/branch-2 Commit: cf550c2be123c91b44b110b5c5ef5060008cc0a7 Parents: 58b0585 Author: Yu LiAuthored: Tue Oct 17 10:24:07 2017 +0800 Committer: Yu Li Committed: Tue Oct 17 10:25:27 2017 +0800 -- .../java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/cf550c2b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java -- diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java index 20b2d42..fdbe6c0 100644 --- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java +++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java @@ -172,7 +172,7 @@ public class HFileOutputFormat2 static final String MULTI_TABLE_HFILEOUTPUTFORMAT_CONF_KEY = "hbase.mapreduce.use.multi.table.hfileoutputformat"; - public static final String STORAGE_POLICY_PROPERTY = "hbase.hstore.storagepolicy"; + public static final String STORAGE_POLICY_PROPERTY = HStore.BLOCK_STORAGE_POLICY_KEY; public static final String STORAGE_POLICY_PROPERTY_CF_PREFIX = STORAGE_POLICY_PROPERTY + "."; @Override
hbase git commit: HBASE-19016 Coordinate storage policy property name for table schema and bulkload
Repository: hbase Updated Branches: refs/heads/master 51489b208 -> be4570fc8 HBASE-19016 Coordinate storage policy property name for table schema and bulkload Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/be4570fc Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/be4570fc Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/be4570fc Branch: refs/heads/master Commit: be4570fc85ad51357fdb9eed3f402a59314ad41d Parents: 51489b20 Author: Yu LiAuthored: Tue Oct 17 10:24:07 2017 +0800 Committer: Yu Li Committed: Tue Oct 17 10:24:07 2017 +0800 -- .../java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/be4570fc/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java -- diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java index 20b2d42..fdbe6c0 100644 --- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java +++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat2.java @@ -172,7 +172,7 @@ public class HFileOutputFormat2 static final String MULTI_TABLE_HFILEOUTPUTFORMAT_CONF_KEY = "hbase.mapreduce.use.multi.table.hfileoutputformat"; - public static final String STORAGE_POLICY_PROPERTY = "hbase.hstore.storagepolicy"; + public static final String STORAGE_POLICY_PROPERTY = HStore.BLOCK_STORAGE_POLICY_KEY; public static final String STORAGE_POLICY_PROPERTY_CF_PREFIX = STORAGE_POLICY_PROPERTY + "."; @Override
[19/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.EnableTableProcedureBiConsumer.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName tableName) { -443
[07/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html index 81d256e..73c1480 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BatchOperation.html @@ -2964,5322 +2964,5321 @@ 2956// This is expensive. 2957if (cell.getTimestamp() == HConstants.LATEST_TIMESTAMP CellUtil.isDeleteType(cell)) { 2958 byte[] qual = CellUtil.cloneQualifier(cell); -2959 if (qual == null) qual = HConstants.EMPTY_BYTE_ARRAY; -2960 -2961 Integer count = kvCount.get(qual); -2962 if (count == null) { -2963kvCount.put(qual, 1); -2964 } else { -2965kvCount.put(qual, count + 1); -2966 } -2967 count = kvCount.get(qual); -2968 -2969 Get get = new Get(CellUtil.cloneRow(cell)); -2970 get.setMaxVersions(count); -2971 get.addColumn(family, qual); -2972 if (coprocessorHost != null) { -2973if (!coprocessorHost.prePrepareTimeStampForDeleteVersion(mutation, cell, -2974byteNow, get)) { -2975 updateDeleteLatestVersionTimeStamp(cell, get, count, byteNow); -2976} -2977 } else { -2978 updateDeleteLatestVersionTimeStamp(cell, get, count, byteNow); -2979 } -2980} else { -2981 CellUtil.updateLatestStamp(cell, byteNow, 0); -2982} -2983 } -2984} -2985 } -2986 -2987 void updateDeleteLatestVersionTimeStamp(Cell cell, Get get, int count, byte[] byteNow) -2988 throws IOException { -2989ListCell result = get(get, false); -2990 -2991if (result.size() count) { -2992 // Nothing to delete -2993 CellUtil.updateLatestStamp(cell, byteNow, 0); -2994 return; -2995} -2996if (result.size() count) { -2997 throw new RuntimeException("Unexpected size: " + result.size()); -2998} -2999Cell getCell = result.get(count - 1); -3000CellUtil.setTimestamp(cell, getCell.getTimestamp()); -3001 } -3002 -3003 @Override -3004 public void put(Put put) throws IOException { -3005checkReadOnly(); -3006 -3007// Do a rough check that we have resources to accept a write. The check is -3008// 'rough' in that between the resource check and the call to obtain a -3009// read lock, resources may run out. For now, the thought is that this -3010// will be extremely rare; we'll deal with it when it happens. -3011checkResources(); -3012 startRegionOperation(Operation.PUT); -3013try { -3014 // All edits for the given row (across all column families) must happen atomically. -3015 doBatchMutate(put); -3016} finally { -3017 closeRegionOperation(Operation.PUT); -3018} -3019 } -3020 -3021 /** -3022 * Struct-like class that tracks the progress of a batch operation, -3023 * accumulating status codes and tracking the index at which processing -3024 * is proceeding. -3025 */ -3026 private abstract static class BatchOperationT { -3027T[] operations; -3028int nextIndexToProcess = 0; -3029OperationStatus[] retCodeDetails; -3030WALEdit[] walEditsFromCoprocessors; -3031 -3032public BatchOperation(T[] operations) { -3033 this.operations = operations; -3034 this.retCodeDetails = new OperationStatus[operations.length]; -3035 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -3036 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -3037} -3038 -3039public abstract Mutation getMutation(int index); -3040public abstract long getNonceGroup(int index); -3041public abstract long getNonce(int index); -3042/** This method is potentially expensive and should only be used for non-replay CP path. */ -3043public abstract Mutation[] getMutationsForCoprocs(); -3044public abstract boolean isInReplay(); -3045public abstract long getReplaySequenceId(); -3046 -3047public boolean isDone() { -3048 return nextIndexToProcess == operations.length; -3049} -3050 } -3051 -3052 private static class MutationBatch extends BatchOperationMutation { -3053private long nonceGroup; -3054private long nonce; -3055public MutationBatch(Mutation[] operations, long nonceGroup, long nonce) { -3056 super(operations); -3057 this.nonceGroup = nonceGroup; -3058 this.nonce = nonce; -3059} -3060 -3061@Override -3062public Mutation getMutation(int index) { -3063 return this.operations[index]; -3064} -3065 -3066@Override -3067
hbase-site git commit: INFRA-10751 Empty commit
Repository: hbase-site Updated Branches: refs/heads/asf-site 47abd8e67 -> 37bebb63e INFRA-10751 Empty commit Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/37bebb63 Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/37bebb63 Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/37bebb63 Branch: refs/heads/asf-site Commit: 37bebb63e3e7eded206fbf00be5cd6e1519231af Parents: 47abd8e Author: jenkinsAuthored: Mon Oct 16 15:16:39 2017 + Committer: jenkins Committed: Mon Oct 16 15:16:39 2017 + -- --
[09/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TruncateTableProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TruncateTableProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TruncateTableProcedureBiConsumer.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TruncateTableProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TruncateTableProcedureBiConsumer.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName tableName) {
[30/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncAdmin.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncAdmin.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncAdmin.html index ad90124..319e418 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncAdmin.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncAdmin.html @@ -151,992 +151,938 @@ 143 CompletableFutureVoid deleteTable(TableName tableName); 144 145 /** -146 * Delete tables matching the passed in pattern and wait on completion. Warning: Use this method -147 * carefully, there is no prompting and the effect is immediate. Consider using -148 * {@link #listTableNames(Optional, boolean) } and -149 * {@link #deleteTable(org.apache.hadoop.hbase.TableName)} -150 * @param pattern The pattern to match table names against -151 * @return Table descriptors for tables that couldn't be deleted. The return value will be wrapped -152 * by a {@link CompletableFuture}. The return HTDs are read-only. -153 */ -154 CompletableFutureListTableDescriptor deleteTables(Pattern pattern); -155 -156 /** -157 * Truncate a table. -158 * @param tableName name of table to truncate -159 * @param preserveSplits True if the splits should be preserved -160 */ -161 CompletableFutureVoid truncateTable(TableName tableName, boolean preserveSplits); -162 -163 /** -164 * Enable a table. The table has to be in disabled state for it to be enabled. -165 * @param tableName name of the table -166 */ -167 CompletableFutureVoid enableTable(TableName tableName); -168 -169 /** -170 * Enable tables matching the passed in pattern. Warning: Use this method carefully, there is no -171 * prompting and the effect is immediate. Consider using {@link #listTables(Optional, boolean)} and -172 * {@link #enableTable(TableName)} -173 * @param pattern The pattern to match table names against -174 * @return Table descriptors for tables that couldn't be enabled. The return value will be wrapped -175 * by a {@link CompletableFuture}. The return HTDs are read-only. -176 */ -177 CompletableFutureListTableDescriptor enableTables(Pattern pattern); -178 -179 /** -180 * Disable a table. The table has to be in enabled state for it to be disabled. -181 * @param tableName +146 * Truncate a table. +147 * @param tableName name of table to truncate +148 * @param preserveSplits True if the splits should be preserved +149 */ +150 CompletableFutureVoid truncateTable(TableName tableName, boolean preserveSplits); +151 +152 /** +153 * Enable a table. The table has to be in disabled state for it to be enabled. +154 * @param tableName name of the table +155 */ +156 CompletableFutureVoid enableTable(TableName tableName); +157 +158 /** +159 * Disable a table. The table has to be in enabled state for it to be disabled. +160 * @param tableName +161 */ +162 CompletableFutureVoid disableTable(TableName tableName); +163 +164 /** +165 * @param tableName name of table to check +166 * @return true if table is on-line. The return value will be wrapped by a +167 * {@link CompletableFuture}. +168 */ +169 CompletableFutureBoolean isTableEnabled(TableName tableName); +170 +171 /** +172 * @param tableName name of table to check +173 * @return true if table is off-line. The return value will be wrapped by a +174 * {@link CompletableFuture}. +175 */ +176 CompletableFutureBoolean isTableDisabled(TableName tableName); +177 +178 /** +179 * @param tableName name of table to check +180 * @return true if all regions of the table are available. The return value will be wrapped by a +181 * {@link CompletableFuture}. 182 */ -183 CompletableFutureVoid disableTable(TableName tableName); -184 -185 /** -186 * Disable tables matching the passed in pattern. Warning: Use this method carefully, there is no -187 * prompting and the effect is immediate. Consider using {@link #listTables(Optional, boolean)} and -188 * {@link #disableTable(TableName)} -189 * @param pattern The pattern to match table names against -190 * @return Table descriptors for tables that couldn't be disabled. The return value will be wrapped by a -191 * {@link CompletableFuture}. The return HTDs are read-only. -192 */ -193 CompletableFutureListTableDescriptor disableTables(Pattern pattern); -194 -195 /** -196 * @param tableName name of table to check -197 * @return true if table is on-line. The return value will be wrapped by a -198 * {@link CompletableFuture}. -199 */ -200 CompletableFutureBoolean isTableEnabled(TableName tableName); -201 -202 /** -203 * @param tableName name of table to check -204 * @return true if table is off-line. The
[21/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteTableProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteTableProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteTableProcedureBiConsumer.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteTableProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteTableProcedureBiConsumer.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName tableName) { -443
[10/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableProcedureBiConsumer.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableProcedureBiConsumer.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName tableName) { -443 CompletableFutureTableDescriptor
[46/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/apidocs/src-html/org/apache/hadoop/hbase/client/AsyncAdmin.html -- diff --git a/apidocs/src-html/org/apache/hadoop/hbase/client/AsyncAdmin.html b/apidocs/src-html/org/apache/hadoop/hbase/client/AsyncAdmin.html index ad90124..319e418 100644 --- a/apidocs/src-html/org/apache/hadoop/hbase/client/AsyncAdmin.html +++ b/apidocs/src-html/org/apache/hadoop/hbase/client/AsyncAdmin.html @@ -151,992 +151,938 @@ 143 CompletableFutureVoid deleteTable(TableName tableName); 144 145 /** -146 * Delete tables matching the passed in pattern and wait on completion. Warning: Use this method -147 * carefully, there is no prompting and the effect is immediate. Consider using -148 * {@link #listTableNames(Optional, boolean) } and -149 * {@link #deleteTable(org.apache.hadoop.hbase.TableName)} -150 * @param pattern The pattern to match table names against -151 * @return Table descriptors for tables that couldn't be deleted. The return value will be wrapped -152 * by a {@link CompletableFuture}. The return HTDs are read-only. -153 */ -154 CompletableFutureListTableDescriptor deleteTables(Pattern pattern); -155 -156 /** -157 * Truncate a table. -158 * @param tableName name of table to truncate -159 * @param preserveSplits True if the splits should be preserved -160 */ -161 CompletableFutureVoid truncateTable(TableName tableName, boolean preserveSplits); -162 -163 /** -164 * Enable a table. The table has to be in disabled state for it to be enabled. -165 * @param tableName name of the table -166 */ -167 CompletableFutureVoid enableTable(TableName tableName); -168 -169 /** -170 * Enable tables matching the passed in pattern. Warning: Use this method carefully, there is no -171 * prompting and the effect is immediate. Consider using {@link #listTables(Optional, boolean)} and -172 * {@link #enableTable(TableName)} -173 * @param pattern The pattern to match table names against -174 * @return Table descriptors for tables that couldn't be enabled. The return value will be wrapped -175 * by a {@link CompletableFuture}. The return HTDs are read-only. -176 */ -177 CompletableFutureListTableDescriptor enableTables(Pattern pattern); -178 -179 /** -180 * Disable a table. The table has to be in enabled state for it to be disabled. -181 * @param tableName +146 * Truncate a table. +147 * @param tableName name of table to truncate +148 * @param preserveSplits True if the splits should be preserved +149 */ +150 CompletableFutureVoid truncateTable(TableName tableName, boolean preserveSplits); +151 +152 /** +153 * Enable a table. The table has to be in disabled state for it to be enabled. +154 * @param tableName name of the table +155 */ +156 CompletableFutureVoid enableTable(TableName tableName); +157 +158 /** +159 * Disable a table. The table has to be in enabled state for it to be disabled. +160 * @param tableName +161 */ +162 CompletableFutureVoid disableTable(TableName tableName); +163 +164 /** +165 * @param tableName name of table to check +166 * @return true if table is on-line. The return value will be wrapped by a +167 * {@link CompletableFuture}. +168 */ +169 CompletableFutureBoolean isTableEnabled(TableName tableName); +170 +171 /** +172 * @param tableName name of table to check +173 * @return true if table is off-line. The return value will be wrapped by a +174 * {@link CompletableFuture}. +175 */ +176 CompletableFutureBoolean isTableDisabled(TableName tableName); +177 +178 /** +179 * @param tableName name of table to check +180 * @return true if all regions of the table are available. The return value will be wrapped by a +181 * {@link CompletableFuture}. 182 */ -183 CompletableFutureVoid disableTable(TableName tableName); -184 -185 /** -186 * Disable tables matching the passed in pattern. Warning: Use this method carefully, there is no -187 * prompting and the effect is immediate. Consider using {@link #listTables(Optional, boolean)} and -188 * {@link #disableTable(TableName)} -189 * @param pattern The pattern to match table names against -190 * @return Table descriptors for tables that couldn't be disabled. The return value will be wrapped by a -191 * {@link CompletableFuture}. The return HTDs are read-only. -192 */ -193 CompletableFutureListTableDescriptor disableTables(Pattern pattern); -194 -195 /** -196 * @param tableName name of table to check -197 * @return true if table is on-line. The return value will be wrapped by a -198 * {@link CompletableFuture}. -199 */ -200 CompletableFutureBoolean isTableEnabled(TableName tableName); -201 -202 /** -203 * @param tableName name of table to check -204 * @return true if table is off-line. The return value will
[39/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/org/apache/hadoop/hbase/client/AsyncAdmin.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/AsyncAdmin.html b/devapidocs/org/apache/hadoop/hbase/client/AsyncAdmin.html index 18a03c2..7c68213 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/AsyncAdmin.html +++ b/devapidocs/org/apache/hadoop/hbase/client/AsyncAdmin.html @@ -18,8 +18,8 @@ catch(err) { } //--> -var methods = {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":18,"i6":6,"i7":6,"i8":6,"i9":6,"i10":38,"i11":18,"i12":6,"i13":18,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":18,"i20":6,"i21":6,"i22":6,"i23":6,"i24":6,"i25":18,"i26":6,"i27":6,"i28":6,"i29":6,"i30":6,"i31":6,"i32":6,"i33":6,"i34":6,"i35":6,"i36":6,"i37":6,"i38":6,"i39":6,"i40":6,"i41":18,"i42":6,"i43":6,"i44":6,"i45":6,"i46":6,"i47":6,"i48":6,"i49":18,"i50":6,"i51":6,"i52":6,"i53":6,"i54":18,"i55":6,"i56":18,"i57":6,"i58":6,"i59":6,"i60":6,"i61":6,"i62":6,"i63":6,"i64":6,"i65":6,"i66":6,"i67":6,"i68":6,"i69":6,"i70":18,"i71":6,"i72":6,"i73":6,"i74":6,"i75":6,"i76":6,"i77":6,"i78":18,"i79":6,"i80":18,"i81":6,"i82":18,"i83":6,"i84":18,"i85":6,"i86":6,"i87":18,"i88":6,"i89":18,"i90":6,"i91":6,"i92":6,"i93":6,"i94":6,"i95":6,"i96":6,"i97":6,"i98":6,"i99":6,"i100":6,"i101":6,"i102":6,"i103":6,"i104":6,"i105":6,"i106":6,"i107":6,"i108":6,"i109":6,"i110":6,"i111":6,"i112":6,"i113":6,"i114":6,"i115":18,"i116":18,"i117":6,"i118":6,"i 119":18,"i120":6,"i121":6,"i122":6,"i123":6,"i124":6,"i125":6,"i126":6,"i127":6,"i128":6}; -var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],16:["t5","Default Methods"],32:["t6","Deprecated Methods"]}; +var methods = {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":18,"i6":6,"i7":6,"i8":6,"i9":6,"i10":18,"i11":6,"i12":18,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":18,"i19":6,"i20":6,"i21":6,"i22":6,"i23":6,"i24":18,"i25":6,"i26":6,"i27":6,"i28":6,"i29":6,"i30":6,"i31":6,"i32":6,"i33":6,"i34":6,"i35":6,"i36":18,"i37":6,"i38":6,"i39":6,"i40":6,"i41":6,"i42":6,"i43":6,"i44":18,"i45":6,"i46":6,"i47":6,"i48":6,"i49":18,"i50":6,"i51":18,"i52":6,"i53":6,"i54":6,"i55":6,"i56":6,"i57":6,"i58":6,"i59":6,"i60":6,"i61":6,"i62":6,"i63":6,"i64":6,"i65":18,"i66":6,"i67":6,"i68":6,"i69":6,"i70":6,"i71":6,"i72":6,"i73":18,"i74":6,"i75":18,"i76":6,"i77":18,"i78":6,"i79":18,"i80":6,"i81":6,"i82":18,"i83":6,"i84":18,"i85":6,"i86":6,"i87":6,"i88":6,"i89":6,"i90":6,"i91":6,"i92":6,"i93":6,"i94":6,"i95":6,"i96":6,"i97":6,"i98":6,"i99":6,"i100":6,"i101":6,"i102":6,"i103":6,"i104":6,"i105":6,"i106":6,"i107":6,"i108":6,"i109":6,"i110":18,"i111":18,"i112":6,"i113":6,"i114":18,"i115":6,"i116":6,"i117":6,"i118":6,"i 119":6,"i120":6,"i121":6,"i122":6,"i123":6}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],16:["t5","Default Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; var tableTab = "tableTab"; @@ -128,7 +128,7 @@ public interface -All MethodsInstance MethodsAbstract MethodsDefault MethodsDeprecated Methods +All MethodsInstance MethodsAbstract MethodsDefault Methods Modifier and Type Method and Description @@ -198,56 +198,45 @@ public interface -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean -closeRegion(byte[]regionName, - http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">OptionalServerNameserverName) -Deprecated. -As of release 2.0.0, this will be removed in HBase 3.0.0 - (https://issues.apache.org/jira/browse/HBASE-18231;>HBASE-18231). - Use unassign(byte[], boolean). - - - - default http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void compact(TableNametableName) Compact a table. - + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void compact(TableNametableName, http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">Optionalbyte[]columnFamily) Compact a column family within a table. - + default
[01/51] [partial] hbase-site git commit: Published site at .
Repository: hbase-site Updated Branches: refs/heads/asf-site 373abd1ff -> 47abd8e67 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html index 81d256e..73c1480 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.ObservedExceptionsInBatch.html @@ -2964,5322 +2964,5321 @@ 2956// This is expensive. 2957if (cell.getTimestamp() == HConstants.LATEST_TIMESTAMP CellUtil.isDeleteType(cell)) { 2958 byte[] qual = CellUtil.cloneQualifier(cell); -2959 if (qual == null) qual = HConstants.EMPTY_BYTE_ARRAY; -2960 -2961 Integer count = kvCount.get(qual); -2962 if (count == null) { -2963kvCount.put(qual, 1); -2964 } else { -2965kvCount.put(qual, count + 1); -2966 } -2967 count = kvCount.get(qual); -2968 -2969 Get get = new Get(CellUtil.cloneRow(cell)); -2970 get.setMaxVersions(count); -2971 get.addColumn(family, qual); -2972 if (coprocessorHost != null) { -2973if (!coprocessorHost.prePrepareTimeStampForDeleteVersion(mutation, cell, -2974byteNow, get)) { -2975 updateDeleteLatestVersionTimeStamp(cell, get, count, byteNow); -2976} -2977 } else { -2978 updateDeleteLatestVersionTimeStamp(cell, get, count, byteNow); -2979 } -2980} else { -2981 CellUtil.updateLatestStamp(cell, byteNow, 0); -2982} -2983 } -2984} -2985 } -2986 -2987 void updateDeleteLatestVersionTimeStamp(Cell cell, Get get, int count, byte[] byteNow) -2988 throws IOException { -2989ListCell result = get(get, false); -2990 -2991if (result.size() count) { -2992 // Nothing to delete -2993 CellUtil.updateLatestStamp(cell, byteNow, 0); -2994 return; -2995} -2996if (result.size() count) { -2997 throw new RuntimeException("Unexpected size: " + result.size()); -2998} -2999Cell getCell = result.get(count - 1); -3000CellUtil.setTimestamp(cell, getCell.getTimestamp()); -3001 } -3002 -3003 @Override -3004 public void put(Put put) throws IOException { -3005checkReadOnly(); -3006 -3007// Do a rough check that we have resources to accept a write. The check is -3008// 'rough' in that between the resource check and the call to obtain a -3009// read lock, resources may run out. For now, the thought is that this -3010// will be extremely rare; we'll deal with it when it happens. -3011checkResources(); -3012 startRegionOperation(Operation.PUT); -3013try { -3014 // All edits for the given row (across all column families) must happen atomically. -3015 doBatchMutate(put); -3016} finally { -3017 closeRegionOperation(Operation.PUT); -3018} -3019 } -3020 -3021 /** -3022 * Struct-like class that tracks the progress of a batch operation, -3023 * accumulating status codes and tracking the index at which processing -3024 * is proceeding. -3025 */ -3026 private abstract static class BatchOperationT { -3027T[] operations; -3028int nextIndexToProcess = 0; -3029OperationStatus[] retCodeDetails; -3030WALEdit[] walEditsFromCoprocessors; -3031 -3032public BatchOperation(T[] operations) { -3033 this.operations = operations; -3034 this.retCodeDetails = new OperationStatus[operations.length]; -3035 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -3036 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -3037} -3038 -3039public abstract Mutation getMutation(int index); -3040public abstract long getNonceGroup(int index); -3041public abstract long getNonce(int index); -3042/** This method is potentially expensive and should only be used for non-replay CP path. */ -3043public abstract Mutation[] getMutationsForCoprocs(); -3044public abstract boolean isInReplay(); -3045public abstract long getReplaySequenceId(); -3046 -3047public boolean isDone() { -3048 return nextIndexToProcess == operations.length; -3049} -3050 } -3051 -3052 private static class MutationBatch extends BatchOperationMutation { -3053private long nonceGroup; -3054private long nonce; -3055public MutationBatch(Mutation[] operations, long nonceGroup, long nonce) { -3056 super(operations); -3057 this.nonceGroup = nonceGroup; -3058 this.nonce = nonce; -3059} -3060 -3061
[42/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/deprecated-list.html -- diff --git a/devapidocs/deprecated-list.html b/devapidocs/deprecated-list.html index 50c2cc6..ecd49ab 100644 --- a/devapidocs/deprecated-list.html +++ b/devapidocs/deprecated-list.html @@ -510,212 +510,205 @@ -org.apache.hadoop.hbase.client.AsyncAdmin.closeRegion(byte[], OptionalServerName) -As of release 2.0.0, this will be removed in HBase 3.0.0 - (https://issues.apache.org/jira/browse/HBASE-18231;>HBASE-18231). - Use AsyncAdmin.unassign(byte[], boolean). - - - org.apache.hadoop.hbase.client.Admin.closeRegion(byte[], String) As of release 2.0.0, this will be removed in HBase 3.0.0. Use Admin.unassign(byte[], boolean). - + org.apache.hadoop.hbase.client.HBaseAdmin.closeRegion(byte[], String) - + org.apache.hadoop.hbase.client.Admin.closeRegion(ServerName, HRegionInfo) As of release 2.0.0, this will be removed in HBase 3.0.0 (https://issues.apache.org/jira/browse/HBASE-18231;>HBASE-18231). Use Admin.unassign(byte[], boolean). - + org.apache.hadoop.hbase.client.HBaseAdmin.closeRegion(ServerName, HRegionInfo) - + org.apache.hadoop.hbase.client.Admin.closeRegion(String, String) As of release 2.0.0, this will be removed in HBase 3.0.0. Use Admin.unassign(byte[], boolean). - + org.apache.hadoop.hbase.client.HBaseAdmin.closeRegion(String, String) - + org.apache.hadoop.hbase.client.Admin.closeRegionWithEncodedRegionName(String, String) As of release 2.0.0, this will be removed in HBase 3.0.0. Use Admin.unassign(byte[], boolean). - + org.apache.hadoop.hbase.client.HBaseAdmin.closeRegionWithEncodedRegionName(String, String) - + org.apache.hadoop.hbase.filter.CompareFilter.compareFamily(CompareFilter.CompareOp, ByteArrayComparable, Cell) Since 2.0.0. Will be removed in 3.0.0. Use CompareFilter.compareFamily(CompareOperator, ByteArrayComparable, Cell) - + org.apache.hadoop.hbase.filter.CompareFilter.compareQualifier(CompareFilter.CompareOp, ByteArrayComparable, Cell) Since 2.0.0. Will be removed in 3.0.0. Use CompareFilter.compareQualifier(CompareOperator, ByteArrayComparable, Cell) - + org.apache.hadoop.hbase.filter.CompareFilter.compareRow(CompareFilter.CompareOp, ByteArrayComparable, Cell) Since 2.0.0. Will be removed in 3.0.0. Use CompareFilter.compareRow(CompareOperator, ByteArrayComparable, Cell) - + org.apache.hadoop.hbase.filter.CompareFilter.compareValue(CompareFilter.CompareOp, ByteArrayComparable, Cell) Since 2.0.0. Will be removed in 3.0.0. Use CompareFilter.compareValue(CompareOperator, ByteArrayComparable, Cell) - + org.apache.hadoop.hbase.HRegionInfo.convert(HBaseProtos.RegionInfo) As of release 2.0.0, this will be removed in HBase 3.0.0 Use toRegionInfo(HBaseProtos.RegionInfo) in org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil. - + org.apache.hadoop.hbase.HRegionInfo.convert(HRegionInfo) As of release 2.0.0, this will be removed in HBase 3.0.0 Use toRegionInfo(org.apache.hadoop.hbase.client.RegionInfo) in org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil. - + org.apache.hadoop.hbase.mapreduce.CellCreator.create(byte[], int, int, byte[], int, int, byte[], int, int, long, byte[], int, int, String) - + org.apache.hadoop.hbase.regionserver.RpcSchedulerFactory.create(Configuration, PriorityFunction) - + org.apache.hadoop.hbase.regionserver.SimpleRpcSchedulerFactory.create(Configuration, PriorityFunction) - + org.apache.hadoop.hbase.regionserver.FifoRpcSchedulerFactory.create(Configuration, PriorityFunction) - + org.apache.hadoop.hbase.coprocessor.ObserverContextImpl.createAndPrepare(E) - + org.apache.hadoop.hbase.CellUtil.createCell(byte[]) As of release 2.0.0, this will be removed in HBase 3.0.0. Use CellBuilder instead - + org.apache.hadoop.hbase.CellUtil.createCell(byte[], byte[]) As of release 2.0.0, this will be removed in HBase 3.0.0. Use CellBuilder instead - + org.apache.hadoop.hbase.CellUtil.createCell(byte[], byte[], byte[]) As of release 2.0.0, this will be removed in HBase 3.0.0. Use CellBuilder instead - + org.apache.hadoop.hbase.CellUtil.createCell(byte[], byte[], byte[], long, byte, byte[]) As of release 2.0.0, this will be removed in HBase 3.0.0. Use CellBuilder instead - + org.apache.hadoop.hbase.CellUtil.createCell(byte[], byte[], byte[], long, byte, byte[], byte[], long) As of release 2.0.0, this will be removed in HBase 3.0.0. Use ExtendedCellBuilder instead - + org.apache.hadoop.hbase.CellUtil.createCell(byte[], byte[], byte[], long, byte, byte[], long) As of release 2.0.0, this will be removed in HBase 3.0.0.
[04/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html index 81d256e..73c1480 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.html @@ -2964,5322 +2964,5321 @@ 2956// This is expensive. 2957if (cell.getTimestamp() == HConstants.LATEST_TIMESTAMP CellUtil.isDeleteType(cell)) { 2958 byte[] qual = CellUtil.cloneQualifier(cell); -2959 if (qual == null) qual = HConstants.EMPTY_BYTE_ARRAY; -2960 -2961 Integer count = kvCount.get(qual); -2962 if (count == null) { -2963kvCount.put(qual, 1); -2964 } else { -2965kvCount.put(qual, count + 1); -2966 } -2967 count = kvCount.get(qual); -2968 -2969 Get get = new Get(CellUtil.cloneRow(cell)); -2970 get.setMaxVersions(count); -2971 get.addColumn(family, qual); -2972 if (coprocessorHost != null) { -2973if (!coprocessorHost.prePrepareTimeStampForDeleteVersion(mutation, cell, -2974byteNow, get)) { -2975 updateDeleteLatestVersionTimeStamp(cell, get, count, byteNow); -2976} -2977 } else { -2978 updateDeleteLatestVersionTimeStamp(cell, get, count, byteNow); -2979 } -2980} else { -2981 CellUtil.updateLatestStamp(cell, byteNow, 0); -2982} -2983 } -2984} -2985 } -2986 -2987 void updateDeleteLatestVersionTimeStamp(Cell cell, Get get, int count, byte[] byteNow) -2988 throws IOException { -2989ListCell result = get(get, false); -2990 -2991if (result.size() count) { -2992 // Nothing to delete -2993 CellUtil.updateLatestStamp(cell, byteNow, 0); -2994 return; -2995} -2996if (result.size() count) { -2997 throw new RuntimeException("Unexpected size: " + result.size()); -2998} -2999Cell getCell = result.get(count - 1); -3000CellUtil.setTimestamp(cell, getCell.getTimestamp()); -3001 } -3002 -3003 @Override -3004 public void put(Put put) throws IOException { -3005checkReadOnly(); -3006 -3007// Do a rough check that we have resources to accept a write. The check is -3008// 'rough' in that between the resource check and the call to obtain a -3009// read lock, resources may run out. For now, the thought is that this -3010// will be extremely rare; we'll deal with it when it happens. -3011checkResources(); -3012 startRegionOperation(Operation.PUT); -3013try { -3014 // All edits for the given row (across all column families) must happen atomically. -3015 doBatchMutate(put); -3016} finally { -3017 closeRegionOperation(Operation.PUT); -3018} -3019 } -3020 -3021 /** -3022 * Struct-like class that tracks the progress of a batch operation, -3023 * accumulating status codes and tracking the index at which processing -3024 * is proceeding. -3025 */ -3026 private abstract static class BatchOperationT { -3027T[] operations; -3028int nextIndexToProcess = 0; -3029OperationStatus[] retCodeDetails; -3030WALEdit[] walEditsFromCoprocessors; -3031 -3032public BatchOperation(T[] operations) { -3033 this.operations = operations; -3034 this.retCodeDetails = new OperationStatus[operations.length]; -3035 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -3036 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -3037} -3038 -3039public abstract Mutation getMutation(int index); -3040public abstract long getNonceGroup(int index); -3041public abstract long getNonce(int index); -3042/** This method is potentially expensive and should only be used for non-replay CP path. */ -3043public abstract Mutation[] getMutationsForCoprocs(); -3044public abstract boolean isInReplay(); -3045public abstract long getReplaySequenceId(); -3046 -3047public boolean isDone() { -3048 return nextIndexToProcess == operations.length; -3049} -3050 } -3051 -3052 private static class MutationBatch extends BatchOperationMutation { -3053private long nonceGroup; -3054private long nonce; -3055public MutationBatch(Mutation[] operations, long nonceGroup, long nonce) { -3056 super(operations); -3057 this.nonceGroup = nonceGroup; -3058 this.nonce = nonce; -3059} -3060 -3061@Override -3062public Mutation getMutation(int index) { -3063 return this.operations[index]; -3064} -3065 -3066@Override -3067public long
[12/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.SplitTableRegionProcedureBiConsumer.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName
[13/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ProcedureBiConsumer.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ProcedureBiConsumer.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName tableName) { -443 CompletableFutureTableDescriptor future = new
[24/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateTableProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateTableProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateTableProcedureBiConsumer.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateTableProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateTableProcedureBiConsumer.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName tableName) { -443
[35/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/org/apache/hadoop/hbase/client/class-use/RawAsyncHBaseAdmin.TableOperator.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/RawAsyncHBaseAdmin.TableOperator.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/RawAsyncHBaseAdmin.TableOperator.html index 2ce4f03..18de7a2 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RawAsyncHBaseAdmin.TableOperator.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RawAsyncHBaseAdmin.TableOperator.html @@ -72,51 +72,7 @@ Uses of Interfaceorg.apache.hadoop.hbase.client.RawAsyncHBaseAdmin.TableOperator - - - - -Packages that use RawAsyncHBaseAdmin.TableOperator - -Package -Description - - - -org.apache.hadoop.hbase.client - -Provides HBase Client - - - - - - - - - - -Uses of RawAsyncHBaseAdmin.TableOperator in org.apache.hadoop.hbase.client - -Methods in org.apache.hadoop.hbase.client with parameters of type RawAsyncHBaseAdmin.TableOperator - -Modifier and Type -Method and Description - - - -private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableDescriptor -RawAsyncHBaseAdmin.batchTableOperations(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Patternpattern, -RawAsyncHBaseAdmin.TableOperatoroperator, -http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringoperationType) - - - - - - - - +No usage of org.apache.hadoop.hbase.client.RawAsyncHBaseAdmin.TableOperator http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/org/apache/hadoop/hbase/client/class-use/TableDescriptor.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/TableDescriptor.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/TableDescriptor.html index 2b92aea..450053c 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/TableDescriptor.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/TableDescriptor.html @@ -470,54 +470,6 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableDescriptor -RawAsyncHBaseAdmin.batchTableOperations(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Patternpattern, -RawAsyncHBaseAdmin.TableOperatoroperator, -http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringoperationType) - - -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableDescriptor -AsyncHBaseAdmin.deleteTables(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Patternpattern) - - -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableDescriptor -AsyncAdmin.deleteTables(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Patternpattern) -Delete tables matching the passed in pattern and wait on completion. - - - -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableDescriptor -RawAsyncHBaseAdmin.deleteTables(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Patternpattern) - -
[40/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html index 94eebdd..b38fd03 100644 --- a/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html +++ b/devapidocs/org/apache/hadoop/hbase/class-use/TableName.html @@ -3075,10 +3075,6 @@ service. byte[]regionName) -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturePairhttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer,http://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer -AsyncHBaseAdmin.getAlterStatus(TableNametableName) - - Pairhttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer,http://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer Admin.getAlterStatus(TableNametableName) Deprecated. @@ -3087,130 +3083,119 @@ service. - + Pairhttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer,http://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer HBaseAdmin.getAlterStatus(TableNametableName) - -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturePairhttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer,http://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer -AsyncAdmin.getAlterStatus(TableNametableName) -Get the status of alter command - indicates how many regions have received the updated schema - Asynchronous operation. - - -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturePairhttp://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer,http://docs.oracle.com/javase/8/docs/api/java/lang/Integer.html?is-external=true; title="class or interface in java.lang">Integer -RawAsyncHBaseAdmin.getAlterStatus(TableNametableName) - - BufferedMutator ConnectionImplementation.getBufferedMutator(TableNametableName) - + default AsyncBufferedMutator AsyncConnection.getBufferedMutator(TableNametableName) Retrieve an AsyncBufferedMutator for performing client-side buffering of writes. - + BufferedMutator Connection.getBufferedMutator(TableNametableName) Retrieve a BufferedMutator for performing client-side buffering of writes. - + default AsyncBufferedMutator AsyncConnection.getBufferedMutator(TableNametableName, http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true; title="class or interface in java.util.concurrent">ExecutorServicepool) Retrieve an AsyncBufferedMutator for performing client-side buffering of writes. - + AsyncBufferedMutatorBuilder AsyncConnection.getBufferedMutatorBuilder(TableNametableName) Returns an AsyncBufferedMutatorBuilder for creating AsyncBufferedMutator. - + AsyncBufferedMutatorBuilder AsyncConnectionImpl.getBufferedMutatorBuilder(TableNametableName) - + AsyncBufferedMutatorBuilder AsyncConnection.getBufferedMutatorBuilder(TableNametableName, http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true; title="class or interface in java.util.concurrent">ExecutorServicepool) Returns an AsyncBufferedMutatorBuilder for creating AsyncBufferedMutator. - + AsyncBufferedMutatorBuilder AsyncConnectionImpl.getBufferedMutatorBuilder(TableNametableName, http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/ExecutorService.html?is-external=true; title="class or interface in java.util.concurrent">ExecutorServicepool) - + RegionLocations MetaCache.getCachedLocation(TableNametableName, byte[]row) Search the cache for a location that fits our table and row key. - + (package private) RegionLocations ConnectionImplementation.getCachedLocation(TableNametableName, byte[]row) Search the cache for a location that fits our table and row key. - +
[36/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html index 9ef2a83..f31343f 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html +++ b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i 109":10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10,"i136":10,"i137":10,"i138":10,"i139":10,"i140":10,"i141":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i 109":10,"i110":10,"i111":10,"i112":10,"i113":10,"i114":10,"i115":10,"i116":10,"i117":10,"i118":10,"i119":10,"i120":10,"i121":10,"i122":10,"i123":10,"i124":10,"i125":10,"i126":10,"i127":10,"i128":10,"i129":10,"i130":10,"i131":10,"i132":10,"i133":10,"i134":10,"i135":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -355,12 +355,6 @@ implements -private http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableDescriptor -batchTableOperations(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Patternpattern, -RawAsyncHBaseAdmin.TableOperatoroperator, -http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringoperationType) - - private PREQ,PRESP,RESPhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureRESP call(HBaseRpcControllercontroller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterService.Interfacestub, @@ -368,45 +362,38 @@ implements RawAsyncHBaseAdmin.MasterRpcCallPRESP,PREQrpcCall, RawAsyncHBaseAdmin.ConverterRESP,PRESPrespConverter) - + private void checkAndGetTableName(byte[]encodeRegionName, http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/atomic/AtomicReference.html?is-external=true; title="class or interface in java.util.concurrent.atomic">AtomicReferenceTableNametableName, http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class
[03/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html index 81d256e..73c1480 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResultImpl.html @@ -2964,5322 +2964,5321 @@ 2956// This is expensive. 2957if (cell.getTimestamp() == HConstants.LATEST_TIMESTAMP CellUtil.isDeleteType(cell)) { 2958 byte[] qual = CellUtil.cloneQualifier(cell); -2959 if (qual == null) qual = HConstants.EMPTY_BYTE_ARRAY; -2960 -2961 Integer count = kvCount.get(qual); -2962 if (count == null) { -2963kvCount.put(qual, 1); -2964 } else { -2965kvCount.put(qual, count + 1); -2966 } -2967 count = kvCount.get(qual); -2968 -2969 Get get = new Get(CellUtil.cloneRow(cell)); -2970 get.setMaxVersions(count); -2971 get.addColumn(family, qual); -2972 if (coprocessorHost != null) { -2973if (!coprocessorHost.prePrepareTimeStampForDeleteVersion(mutation, cell, -2974byteNow, get)) { -2975 updateDeleteLatestVersionTimeStamp(cell, get, count, byteNow); -2976} -2977 } else { -2978 updateDeleteLatestVersionTimeStamp(cell, get, count, byteNow); -2979 } -2980} else { -2981 CellUtil.updateLatestStamp(cell, byteNow, 0); -2982} -2983 } -2984} -2985 } -2986 -2987 void updateDeleteLatestVersionTimeStamp(Cell cell, Get get, int count, byte[] byteNow) -2988 throws IOException { -2989ListCell result = get(get, false); -2990 -2991if (result.size() count) { -2992 // Nothing to delete -2993 CellUtil.updateLatestStamp(cell, byteNow, 0); -2994 return; -2995} -2996if (result.size() count) { -2997 throw new RuntimeException("Unexpected size: " + result.size()); -2998} -2999Cell getCell = result.get(count - 1); -3000CellUtil.setTimestamp(cell, getCell.getTimestamp()); -3001 } -3002 -3003 @Override -3004 public void put(Put put) throws IOException { -3005checkReadOnly(); -3006 -3007// Do a rough check that we have resources to accept a write. The check is -3008// 'rough' in that between the resource check and the call to obtain a -3009// read lock, resources may run out. For now, the thought is that this -3010// will be extremely rare; we'll deal with it when it happens. -3011checkResources(); -3012 startRegionOperation(Operation.PUT); -3013try { -3014 // All edits for the given row (across all column families) must happen atomically. -3015 doBatchMutate(put); -3016} finally { -3017 closeRegionOperation(Operation.PUT); -3018} -3019 } -3020 -3021 /** -3022 * Struct-like class that tracks the progress of a batch operation, -3023 * accumulating status codes and tracking the index at which processing -3024 * is proceeding. -3025 */ -3026 private abstract static class BatchOperationT { -3027T[] operations; -3028int nextIndexToProcess = 0; -3029OperationStatus[] retCodeDetails; -3030WALEdit[] walEditsFromCoprocessors; -3031 -3032public BatchOperation(T[] operations) { -3033 this.operations = operations; -3034 this.retCodeDetails = new OperationStatus[operations.length]; -3035 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -3036 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -3037} -3038 -3039public abstract Mutation getMutation(int index); -3040public abstract long getNonceGroup(int index); -3041public abstract long getNonce(int index); -3042/** This method is potentially expensive and should only be used for non-replay CP path. */ -3043public abstract Mutation[] getMutationsForCoprocs(); -3044public abstract boolean isInReplay(); -3045public abstract long getReplaySequenceId(); -3046 -3047public boolean isDone() { -3048 return nextIndexToProcess == operations.length; -3049} -3050 } -3051 -3052 private static class MutationBatch extends BatchOperationMutation { -3053private long nonceGroup; -3054private long nonce; -3055public MutationBatch(Mutation[] operations, long nonceGroup, long nonce) { -3056 super(operations); -3057 this.nonceGroup = nonceGroup; -3058 this.nonce = nonce; -3059} -3060 -3061@Override -3062public Mutation getMutation(int index) { -3063 return this.operations[index]; -3064} -3065 -3066@Override
[27/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AdminRpcCall.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AdminRpcCall.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AdminRpcCall.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AdminRpcCall.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AdminRpcCall.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName tableName) { -443 CompletableFutureTableDescriptor future = new CompletableFuture(); -444
[26/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.Converter.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.Converter.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.Converter.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.Converter.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.Converter.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName tableName) { -443 CompletableFutureTableDescriptor future = new CompletableFuture(); -444this.ListTableSchema
[37/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html index 8e6a3a4..ede40d1 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html +++ b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html @@ -127,7 +127,7 @@ var activeTableTab = "activeTableTab"; -private class RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer +private class RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer extends RawAsyncHBaseAdmin.TableProcedureBiConsumer @@ -240,7 +240,7 @@ extends AddColumnFamilyProcedureBiConsumer -AddColumnFamilyProcedureBiConsumer(AsyncAdminadmin, +AddColumnFamilyProcedureBiConsumer(AsyncAdminadmin, TableNametableName) @@ -258,7 +258,7 @@ extends getOperationType -http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetOperationType() +http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetOperationType() Specified by: getOperationTypein classRawAsyncHBaseAdmin.TableProcedureBiConsumer http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html index 023a01b..8bf8e6c 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html +++ b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html @@ -127,7 +127,7 @@ var activeTableTab = "activeTableTab"; -private class RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer +private class RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer extends RawAsyncHBaseAdmin.NamespaceProcedureBiConsumer @@ -240,7 +240,7 @@ extends CreateNamespaceProcedureBiConsumer -CreateNamespaceProcedureBiConsumer(AsyncAdminadmin, +CreateNamespaceProcedureBiConsumer(AsyncAdminadmin, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringnamespaceName) @@ -258,7 +258,7 @@ extends getOperationType -http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetOperationType() +http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetOperationType() Specified by: getOperationTypein classRawAsyncHBaseAdmin.NamespaceProcedureBiConsumer http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateTableProcedureBiConsumer.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateTableProcedureBiConsumer.html b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateTableProcedureBiConsumer.html index 26588f2..9ccbcc1 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateTableProcedureBiConsumer.html +++ b/devapidocs/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateTableProcedureBiConsumer.html @@ -127,7 +127,7 @@ var activeTableTab = "activeTableTab"; -private class RawAsyncHBaseAdmin.CreateTableProcedureBiConsumer +private class RawAsyncHBaseAdmin.CreateTableProcedureBiConsumer extends RawAsyncHBaseAdmin.TableProcedureBiConsumer @@ -240,7 +240,7 @@ extends CreateTableProcedureBiConsumer -CreateTableProcedureBiConsumer(AsyncAdminadmin, +CreateTableProcedureBiConsumer(AsyncAdminadmin, TableNametableName) @@ -258,7 +258,7 @@ extends getOperationType -http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetOperationType() +http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringgetOperationType() Specified by: getOperationTypein classRawAsyncHBaseAdmin.TableProcedureBiConsumer
[41/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/index-all.html -- diff --git a/devapidocs/index-all.html b/devapidocs/index-all.html index 31f67b1..7eb5dff 100644 --- a/devapidocs/index-all.html +++ b/devapidocs/index-all.html @@ -5655,8 +5655,6 @@ batchStartKey - Variable in class org.apache.hadoop.hbase.mapreduce.HashTable.ResultHasher -batchTableOperations(Pattern, RawAsyncHBaseAdmin.TableOperator, String) - Method in class org.apache.hadoop.hbase.client.RawAsyncHBaseAdmin - bcn - Variable in class org.apache.hadoop.hbase.tmpl.regionserver.BlockCacheViewTmplImpl bcn - Variable in class org.apache.hadoop.hbase.tmpl.regionserver.RSStatusTmpl @@ -12880,16 +12878,6 @@ Use Admin.unassign(byte[], boolean). -closeRegion(byte[], OptionalServerName) - Method in interface org.apache.hadoop.hbase.client.AsyncAdmin - -Deprecated. -As of release 2.0.0, this will be removed in HBase 3.0.0 - (https://issues.apache.org/jira/browse/HBASE-18231;>HBASE-18231). - Use AsyncAdmin.unassign(byte[], boolean). - - -closeRegion(byte[], OptionalServerName) - Method in class org.apache.hadoop.hbase.client.AsyncHBaseAdmin - closeRegion(String, String) - Method in class org.apache.hadoop.hbase.client.HBaseAdmin Deprecated. @@ -12902,8 +12890,6 @@ Deprecated. -closeRegion(byte[], OptionalServerName) - Method in class org.apache.hadoop.hbase.client.RawAsyncHBaseAdmin - closeRegion(String, boolean, ServerName) - Method in class org.apache.hadoop.hbase.regionserver.HRegionServer Close asynchronously a region, can be called from the master or internally by the regionserver @@ -24076,20 +24062,12 @@ and Admin.deleteTable(TableName) -deleteTables(Pattern) - Method in interface org.apache.hadoop.hbase.client.AsyncAdmin - -Delete tables matching the passed in pattern and wait on completion. - -deleteTables(Pattern) - Method in class org.apache.hadoop.hbase.client.AsyncHBaseAdmin - deleteTables(String) - Method in class org.apache.hadoop.hbase.client.HBaseAdmin deleteTables(Pattern) - Method in class org.apache.hadoop.hbase.client.HBaseAdmin Delete tables matching the passed in pattern and wait on completion. -deleteTables(Pattern) - Method in class org.apache.hadoop.hbase.client.RawAsyncHBaseAdmin - deleteTableSnapshots(String, String) - Method in interface org.apache.hadoop.hbase.client.Admin Deprecated. @@ -24880,18 +24858,10 @@ and Admin.disableTable(org.apache.hadoop.hbase.TableName) -disableTables(Pattern) - Method in interface org.apache.hadoop.hbase.client.AsyncAdmin - -Disable tables matching the passed in pattern. - -disableTables(Pattern) - Method in class org.apache.hadoop.hbase.client.AsyncHBaseAdmin - disableTables(String) - Method in class org.apache.hadoop.hbase.client.HBaseAdmin disableTables(Pattern) - Method in class org.apache.hadoop.hbase.client.HBaseAdmin -disableTables(Pattern) - Method in class org.apache.hadoop.hbase.client.RawAsyncHBaseAdmin - DisableTableViolationPolicyEnforcement - Class in org.apache.hadoop.hbase.quotas.policies A SpaceViolationPolicyEnforcement which disables the table. @@ -26258,18 +26228,10 @@ and Admin.enableTable(org.apache.hadoop.hbase.TableName) -enableTables(Pattern) - Method in interface org.apache.hadoop.hbase.client.AsyncAdmin - -Enable tables matching the passed in pattern. - -enableTables(Pattern) - Method in class org.apache.hadoop.hbase.client.AsyncHBaseAdmin - enableTables(String) - Method in class org.apache.hadoop.hbase.client.HBaseAdmin enableTables(Pattern) - Method in class org.apache.hadoop.hbase.client.HBaseAdmin -enableTables(Pattern) - Method in class org.apache.hadoop.hbase.client.RawAsyncHBaseAdmin - enableTestingStats() - Method in class org.apache.hadoop.hbase.io.hfile.CompoundBloomFilter enableValues() - Method in class org.apache.hadoop.hbase.wal.WALPrettyPrinter @@ -33260,19 +33222,10 @@ on an operation. -getAlterStatus(TableName) - Method in interface org.apache.hadoop.hbase.client.AsyncAdmin - -Get the status of alter command - indicates how many regions have received the updated schema - Asynchronous operation. - -getAlterStatus(TableName) - Method in class org.apache.hadoop.hbase.client.AsyncHBaseAdmin - getAlterStatus(TableName) - Method in class org.apache.hadoop.hbase.client.HBaseAdmin getAlterStatus(byte[]) - Method in class org.apache.hadoop.hbase.client.HBaseAdmin -getAlterStatus(TableName) - Method in class org.apache.hadoop.hbase.client.RawAsyncHBaseAdmin - getAncestors(BackupInfo) - Method in class org.apache.hadoop.hbase.backup.impl.BackupManager Get direct ancestors of the current backup. http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/org/apache/hadoop/hbase/backup/package-tree.html
[47/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/apidocs/org/apache/hadoop/hbase/client/class-use/TableDescriptor.html -- diff --git a/apidocs/org/apache/hadoop/hbase/client/class-use/TableDescriptor.html b/apidocs/org/apache/hadoop/hbase/client/class-use/TableDescriptor.html index 81e4e99..4992e36 100644 --- a/apidocs/org/apache/hadoop/hbase/client/class-use/TableDescriptor.html +++ b/apidocs/org/apache/hadoop/hbase/client/class-use/TableDescriptor.html @@ -224,67 +224,49 @@ Input/OutputFormats, a table indexing MapReduce job, and utility methods. -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableDescriptor -AsyncAdmin.deleteTables(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Patternpattern) -Delete tables matching the passed in pattern and wait on completion. - - - -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableDescriptor -AsyncAdmin.disableTables(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Patternpattern) -Disable tables matching the passed in pattern. - - - -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableDescriptor -AsyncAdmin.enableTables(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Patternpattern) -Enable tables matching the passed in pattern. - - - http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureTableDescriptor AsyncAdmin.getTableDescriptor(TableNametableName) Method for getting the tableDescriptor - + http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableDescriptor Admin.listTableDescriptors() List all the userspace tables. - + http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableDescriptor Admin.listTableDescriptors(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableNametableNames) Get tableDescriptors. - + http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableDescriptor Admin.listTableDescriptors(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Patternpattern) List all the userspace tables that match the given pattern. - + http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableDescriptor Admin.listTableDescriptors(http://docs.oracle.com/javase/8/docs/api/java/util/regex/Pattern.html?is-external=true; title="class or interface in java.util.regex">Patternpattern, booleanincludeSysTables) List all the tables matching the given pattern. - + http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableDescriptor Admin.listTableDescriptorsByNamespace(byte[]name) Get list of table descriptors by namespace. - + default http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableDescriptor AsyncAdmin.listTables() List all the userspace tables. - + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListTableDescriptor
[38/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html b/devapidocs/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html index 1e2cb92..63a1e3f 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html +++ b/devapidocs/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html @@ -18,7 +18,7 @@ catch(err) { } //--> -var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10,"i106":10,"i107":10,"i108":10,"i 109":10,"i110":10}; +var methods = {"i0":10,"i1":10,"i2":10,"i3":10,"i4":10,"i5":10,"i6":10,"i7":10,"i8":10,"i9":10,"i10":10,"i11":10,"i12":10,"i13":10,"i14":10,"i15":10,"i16":10,"i17":10,"i18":10,"i19":10,"i20":10,"i21":10,"i22":10,"i23":10,"i24":10,"i25":10,"i26":10,"i27":10,"i28":10,"i29":10,"i30":10,"i31":10,"i32":10,"i33":10,"i34":10,"i35":10,"i36":10,"i37":10,"i38":10,"i39":10,"i40":10,"i41":10,"i42":10,"i43":10,"i44":10,"i45":10,"i46":10,"i47":10,"i48":10,"i49":10,"i50":10,"i51":10,"i52":10,"i53":10,"i54":10,"i55":10,"i56":10,"i57":10,"i58":10,"i59":10,"i60":10,"i61":10,"i62":10,"i63":10,"i64":10,"i65":10,"i66":10,"i67":10,"i68":10,"i69":10,"i70":10,"i71":10,"i72":10,"i73":10,"i74":10,"i75":10,"i76":10,"i77":10,"i78":10,"i79":10,"i80":10,"i81":10,"i82":10,"i83":10,"i84":10,"i85":10,"i86":10,"i87":10,"i88":10,"i89":10,"i90":10,"i91":10,"i92":10,"i93":10,"i94":10,"i95":10,"i96":10,"i97":10,"i98":10,"i99":10,"i100":10,"i101":10,"i102":10,"i103":10,"i104":10,"i105":10}; var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],8:["t4","Concrete Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; @@ -244,40 +244,33 @@ implements -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean -closeRegion(byte[]regionName, - http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">OptionalServerNameunused) -Close a region. - - - http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void compact(TableNametableName, http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">Optionalbyte[]columnFamily) Compact a column family within a table. - + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void compactRegion(byte[]regionName, http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">Optionalbyte[]columnFamily) Compact a column family within a region. - + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void compactRegionServer(ServerNameserverName) Compact all regions on the region server. - + S,Rhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFutureR
[25/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.CreateNamespaceProcedureBiConsumer.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName
[31/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html index e74c228..b981ff8 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/package-tree.html @@ -683,19 +683,19 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.regionserver.ScannerContext.NextState -org.apache.hadoop.hbase.regionserver.BloomType -org.apache.hadoop.hbase.regionserver.TimeRangeTracker.Type -org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage -org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status -org.apache.hadoop.hbase.regionserver.ScanType org.apache.hadoop.hbase.regionserver.HRegion.FlushResult.Result -org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType -org.apache.hadoop.hbase.regionserver.Region.Operation +org.apache.hadoop.hbase.regionserver.ScanType org.apache.hadoop.hbase.regionserver.ScannerContext.LimitScope +org.apache.hadoop.hbase.regionserver.MetricsRegionServerSourceFactoryImpl.FactoryStorage +org.apache.hadoop.hbase.regionserver.CompactingMemStore.IndexType org.apache.hadoop.hbase.regionserver.MemStoreCompactor.Action org.apache.hadoop.hbase.regionserver.FlushType +org.apache.hadoop.hbase.regionserver.Region.Operation org.apache.hadoop.hbase.regionserver.DefaultHeapMemoryTuner.StepDirection +org.apache.hadoop.hbase.regionserver.SplitLogWorker.TaskExecutor.Status +org.apache.hadoop.hbase.regionserver.TimeRangeTracker.Type +org.apache.hadoop.hbase.regionserver.ScannerContext.NextState +org.apache.hadoop.hbase.regionserver.BloomType http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html index 9b7ac1f..9d29552 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/querymatcher/package-tree.html @@ -130,9 +130,9 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteResult org.apache.hadoop.hbase.regionserver.querymatcher.ScanQueryMatcher.MatchCode org.apache.hadoop.hbase.regionserver.querymatcher.StripeCompactionScanQueryMatcher.DropDeletesInOutput +org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker.DeleteResult http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html index 870907a..b2572fc 100644 --- a/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html +++ b/devapidocs/org/apache/hadoop/hbase/replication/regionserver/package-tree.html @@ -202,8 +202,8 @@ java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Enum.html?is-external=true; title="class or interface in java.lang">EnumE (implements java.lang.http://docs.oracle.com/javase/8/docs/api/java/lang/Comparable.html?is-external=true; title="class or interface in java.lang">ComparableT, java.io.http://docs.oracle.com/javase/8/docs/api/java/io/Serializable.html?is-external=true; title="class or interface in java.io">Serializable) -org.apache.hadoop.hbase.replication.regionserver.MetricsReplicationSourceFactoryImpl.SourceHolder org.apache.hadoop.hbase.replication.regionserver.ReplicationSourceShipper.WorkerState
[02/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html index 81d256e..73c1480 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html @@ -2964,5322 +2964,5321 @@ 2956// This is expensive. 2957if (cell.getTimestamp() == HConstants.LATEST_TIMESTAMP CellUtil.isDeleteType(cell)) { 2958 byte[] qual = CellUtil.cloneQualifier(cell); -2959 if (qual == null) qual = HConstants.EMPTY_BYTE_ARRAY; -2960 -2961 Integer count = kvCount.get(qual); -2962 if (count == null) { -2963kvCount.put(qual, 1); -2964 } else { -2965kvCount.put(qual, count + 1); -2966 } -2967 count = kvCount.get(qual); -2968 -2969 Get get = new Get(CellUtil.cloneRow(cell)); -2970 get.setMaxVersions(count); -2971 get.addColumn(family, qual); -2972 if (coprocessorHost != null) { -2973if (!coprocessorHost.prePrepareTimeStampForDeleteVersion(mutation, cell, -2974byteNow, get)) { -2975 updateDeleteLatestVersionTimeStamp(cell, get, count, byteNow); -2976} -2977 } else { -2978 updateDeleteLatestVersionTimeStamp(cell, get, count, byteNow); -2979 } -2980} else { -2981 CellUtil.updateLatestStamp(cell, byteNow, 0); -2982} -2983 } -2984} -2985 } -2986 -2987 void updateDeleteLatestVersionTimeStamp(Cell cell, Get get, int count, byte[] byteNow) -2988 throws IOException { -2989ListCell result = get(get, false); -2990 -2991if (result.size() count) { -2992 // Nothing to delete -2993 CellUtil.updateLatestStamp(cell, byteNow, 0); -2994 return; -2995} -2996if (result.size() count) { -2997 throw new RuntimeException("Unexpected size: " + result.size()); -2998} -2999Cell getCell = result.get(count - 1); -3000CellUtil.setTimestamp(cell, getCell.getTimestamp()); -3001 } -3002 -3003 @Override -3004 public void put(Put put) throws IOException { -3005checkReadOnly(); -3006 -3007// Do a rough check that we have resources to accept a write. The check is -3008// 'rough' in that between the resource check and the call to obtain a -3009// read lock, resources may run out. For now, the thought is that this -3010// will be extremely rare; we'll deal with it when it happens. -3011checkResources(); -3012 startRegionOperation(Operation.PUT); -3013try { -3014 // All edits for the given row (across all column families) must happen atomically. -3015 doBatchMutate(put); -3016} finally { -3017 closeRegionOperation(Operation.PUT); -3018} -3019 } -3020 -3021 /** -3022 * Struct-like class that tracks the progress of a batch operation, -3023 * accumulating status codes and tracking the index at which processing -3024 * is proceeding. -3025 */ -3026 private abstract static class BatchOperationT { -3027T[] operations; -3028int nextIndexToProcess = 0; -3029OperationStatus[] retCodeDetails; -3030WALEdit[] walEditsFromCoprocessors; -3031 -3032public BatchOperation(T[] operations) { -3033 this.operations = operations; -3034 this.retCodeDetails = new OperationStatus[operations.length]; -3035 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -3036 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -3037} -3038 -3039public abstract Mutation getMutation(int index); -3040public abstract long getNonceGroup(int index); -3041public abstract long getNonce(int index); -3042/** This method is potentially expensive and should only be used for non-replay CP path. */ -3043public abstract Mutation[] getMutationsForCoprocs(); -3044public abstract boolean isInReplay(); -3045public abstract long getReplaySequenceId(); -3046 -3047public boolean isDone() { -3048 return nextIndexToProcess == operations.length; -3049} -3050 } -3051 -3052 private static class MutationBatch extends BatchOperationMutation { -3053private long nonceGroup; -3054private long nonce; -3055public MutationBatch(Mutation[] operations, long nonceGroup, long nonce) { -3056 super(operations); -3057 this.nonceGroup = nonceGroup; -3058 this.nonce = nonce; -3059} -3060 -3061@Override -3062public Mutation getMutation(int index) { -3063 return this.operations[index]; -3064} -3065 -3066@Override -3067
[43/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/checkstyle.rss -- diff --git a/checkstyle.rss b/checkstyle.rss index ced9295..2cdc726 100644 --- a/checkstyle.rss +++ b/checkstyle.rss @@ -26,7 +26,7 @@ under the License. 2007 - 2017 The Apache Software Foundation File: 2055, - Errors: 13608, + Errors: 13609, Warnings: 0, Infos: 0 @@ -8245,7 +8245,7 @@ under the License. 0 - 127 + 130 @@ -9589,7 +9589,7 @@ under the License. 0 - 2 + 3 @@ -18325,7 +18325,7 @@ under the License. 0 - 18 + 16 @@ -26081,7 +26081,7 @@ under the License. 0 - 228 + 227 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/coc.html -- diff --git a/coc.html b/coc.html index ec1fc8c..303fd09 100644 --- a/coc.html +++ b/coc.html @@ -7,7 +7,7 @@ - + Apache HBase Code of Conduct Policy @@ -380,7 +380,7 @@ email to mailto:priv...@hbase.apache.org;>the priv https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-10-15 + Last Published: 2017-10-16 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/cygwin.html -- diff --git a/cygwin.html b/cygwin.html index f23f27e..64824ce 100644 --- a/cygwin.html +++ b/cygwin.html @@ -7,7 +7,7 @@ - + Apache HBase Installing Apache HBase (TM) on Windows using Cygwin @@ -679,7 +679,7 @@ Now your HBase server is running, start coding and build that next https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-10-15 + Last Published: 2017-10-16 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/dependencies.html -- diff --git a/dependencies.html b/dependencies.html index c282b56..9e8b4cb 100644 --- a/dependencies.html +++ b/dependencies.html @@ -7,7 +7,7 @@ - + Apache HBase Project Dependencies @@ -445,7 +445,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-10-15 + Last Published: 2017-10-16 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/dependency-convergence.html -- diff --git a/dependency-convergence.html b/dependency-convergence.html index dbef118..0a3e5ef 100644 --- a/dependency-convergence.html +++ b/dependency-convergence.html @@ -7,7 +7,7 @@ - + Apache HBase Reactor Dependency Convergence @@ -820,7 +820,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-10-15 + Last Published: 2017-10-16 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/dependency-info.html -- diff --git a/dependency-info.html b/dependency-info.html index b2ead6a..3993787 100644 --- a/dependency-info.html +++ b/dependency-info.html @@ -7,7 +7,7 @@ - + Apache HBase Dependency Information @@ -318,7 +318,7 @@ https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-10-15 + Last Published: 2017-10-16 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/dependency-management.html --
[32/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html index 58bc57a..7bda4ab 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.html @@ -3206,7 +3206,7 @@ implements FIXED_OVERHEAD -public static finallong FIXED_OVERHEAD +public static finallong FIXED_OVERHEAD @@ -3215,7 +3215,7 @@ implements DEEP_OVERHEAD -public static finallong DEEP_OVERHEAD +public static finallong DEEP_OVERHEAD @@ -3224,7 +3224,7 @@ implements MOCKED_LIST -private static finalhttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListCell MOCKED_LIST +private static finalhttp://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListCell MOCKED_LIST A mocked list implementation - discards all updates. @@ -4904,7 +4904,7 @@ publiclong updateDeleteLatestVersionTimeStamp -voidupdateDeleteLatestVersionTimeStamp(Cellcell, +voidupdateDeleteLatestVersionTimeStamp(Cellcell, Getget, intcount, byte[]byteNow) @@ -4921,7 +4921,7 @@ publiclong put -publicvoidput(Putput) +publicvoidput(Putput) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Description copied from interface:Region Puts some data in the table. @@ -4939,7 +4939,7 @@ publiclong batchMutate -publicOperationStatus[]batchMutate(Mutation[]mutations, +publicOperationStatus[]batchMutate(Mutation[]mutations, longnonceGroup, longnonce) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException @@ -4955,7 +4955,7 @@ publiclong batchMutate -publicOperationStatus[]batchMutate(Mutation[]mutations) +publicOperationStatus[]batchMutate(Mutation[]mutations) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Description copied from interface:Region Perform a batch of mutations. @@ -4980,7 +4980,7 @@ publiclong batchReplay -publicOperationStatus[]batchReplay(WALSplitter.MutationReplay[]mutations, +publicOperationStatus[]batchReplay(WALSplitter.MutationReplay[]mutations, longreplaySeqId) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException @@ -4995,7 +4995,7 @@ publiclong batchMutate -OperationStatus[]batchMutate(HRegion.BatchOperation?batchOp) +OperationStatus[]batchMutate(HRegion.BatchOperation?batchOp) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Perform a batch of mutations. It supports only Put and Delete mutations and will ignore other types passed. @@ -5016,7 +5016,7 @@ publiclong doPreBatchMutateHook -privatevoiddoPreBatchMutateHook(HRegion.BatchOperation?batchOp) +privatevoiddoPreBatchMutateHook(HRegion.BatchOperation?batchOp) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Throws: @@ -5030,7 +5030,7 @@ publiclong doMiniBatchMutate -privatevoiddoMiniBatchMutate(HRegion.BatchOperation?batchOp) +privatevoiddoMiniBatchMutate(HRegion.BatchOperation?batchOp) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Called to do a piece of the batch that came in to batchMutate(Mutation[], long, long) In here we also handle replay of edits on region recover. @@ -5046,7 +5046,7 @@ publiclong mergeFamilyMaps -privatevoidmergeFamilyMaps(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in java.util">Mapbyte[],http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListCellfamilyMap, +privatevoidmergeFamilyMaps(http://docs.oracle.com/javase/8/docs/api/java/util/Map.html?is-external=true; title="class or interface in
[51/51] [partial] hbase-site git commit: Published site at .
Published site at . Project: http://git-wip-us.apache.org/repos/asf/hbase-site/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase-site/commit/47abd8e6 Tree: http://git-wip-us.apache.org/repos/asf/hbase-site/tree/47abd8e6 Diff: http://git-wip-us.apache.org/repos/asf/hbase-site/diff/47abd8e6 Branch: refs/heads/asf-site Commit: 47abd8e67f910651fb5415792ec26d8f812eaec2 Parents: 373abd1 Author: jenkinsAuthored: Mon Oct 16 15:16:03 2017 + Committer: jenkins Committed: Mon Oct 16 15:16:03 2017 + -- acid-semantics.html | 4 +- apache_hbase_reference_guide.pdf| 4 +- apidocs/deprecated-list.html| 365 +- apidocs/index-all.html |25 - .../hadoop/hbase/class-use/ServerName.html |19 +- .../hadoop/hbase/class-use/TableName.html | 135 +- .../apache/hadoop/hbase/client/AsyncAdmin.html | 596 +- .../hbase/client/class-use/TableDescriptor.html |32 +- .../hadoop/hbase/util/class-use/Pair.html |16 - .../apache/hadoop/hbase/client/AsyncAdmin.html | 1846 +-- book.html | 2 +- bulk-loads.html | 4 +- checkstyle-aggregate.html | 14474 + checkstyle.rss |10 +- coc.html| 4 +- cygwin.html | 4 +- dependencies.html | 4 +- dependency-convergence.html | 4 +- dependency-info.html| 4 +- dependency-management.html | 4 +- devapidocs/constant-values.html | 6 +- devapidocs/deprecated-list.html | 515 +- devapidocs/index-all.html |47 - .../hadoop/hbase/backup/package-tree.html | 4 +- .../hadoop/hbase/class-use/ServerName.html |51 +- .../hadoop/hbase/class-use/TableName.html | 405 +- .../apache/hadoop/hbase/client/AsyncAdmin.html | 596 +- .../hadoop/hbase/client/AsyncHBaseAdmin.html| 533 +- ...dmin.AddColumnFamilyProcedureBiConsumer.html | 6 +- ...dmin.CreateNamespaceProcedureBiConsumer.html | 6 +- ...aseAdmin.CreateTableProcedureBiConsumer.html | 6 +- ...n.DeleteColumnFamilyProcedureBiConsumer.html | 6 +- ...dmin.DeleteNamespaceProcedureBiConsumer.html | 6 +- ...aseAdmin.DeleteTableProcedureBiConsumer.html | 8 +- ...seAdmin.DisableTableProcedureBiConsumer.html | 6 +- ...aseAdmin.EnableTableProcedureBiConsumer.html | 6 +- ...min.MergeTableRegionProcedureBiConsumer.html | 6 +- ...n.ModifyColumnFamilyProcedureBiConsumer.html | 6 +- ...dmin.ModifyNamespaceProcedureBiConsumer.html | 6 +- ...HBaseAdmin.NamespaceProcedureBiConsumer.html |14 +- .../RawAsyncHBaseAdmin.ProcedureBiConsumer.html |12 +- ...min.SplitTableRegionProcedureBiConsumer.html | 6 +- ...syncHBaseAdmin.TableProcedureBiConsumer.html |14 +- ...eAdmin.TruncateTableProcedureBiConsumer.html | 6 +- .../hadoop/hbase/client/RawAsyncHBaseAdmin.html | 680 +- .../RawAsyncHBaseAdmin.TableOperator.html |46 +- .../hbase/client/class-use/TableDescriptor.html |48 - .../hadoop/hbase/client/package-tree.html |26 +- .../apache/hadoop/hbase/client/package-use.html | 127 +- .../hadoop/hbase/executor/package-tree.html | 2 +- .../hadoop/hbase/filter/package-tree.html | 8 +- .../hadoop/hbase/io/hfile/package-tree.html | 4 +- .../apache/hadoop/hbase/ipc/package-tree.html | 2 +- .../hadoop/hbase/mapreduce/package-tree.html| 4 +- .../hbase/master/balancer/package-tree.html | 2 +- .../hadoop/hbase/master/package-tree.html | 4 +- .../hbase/master/procedure/package-tree.html| 2 +- .../org/apache/hadoop/hbase/package-tree.html |18 +- .../hadoop/hbase/procedure2/package-tree.html | 2 +- .../hadoop/hbase/quotas/package-tree.html | 8 +- .../regionserver/HRegion.BatchOperation.html|26 +- .../regionserver/HRegion.BulkLoadListener.html | 8 +- .../regionserver/HRegion.MutationBatch.html |20 +- .../regionserver/HRegion.RegionScannerImpl.html |90 +- .../hbase/regionserver/HRegion.ReplayBatch.html |18 +- .../regionserver/HRegion.RowLockContext.html|28 +- .../hbase/regionserver/HRegion.RowLockImpl.html |16 +- .../hadoop/hbase/regionserver/HRegion.html | 350 +- .../hadoop/hbase/regionserver/package-tree.html |16 +- .../regionserver/querymatcher/package-tree.html | 2 +- .../replication/regionserver/package-tree.html | 2 +- .../hadoop/hbase/rest/model/package-tree.html | 2 +-
[06/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html index 81d256e..73c1480 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html @@ -2964,5322 +2964,5321 @@ 2956// This is expensive. 2957if (cell.getTimestamp() == HConstants.LATEST_TIMESTAMP CellUtil.isDeleteType(cell)) { 2958 byte[] qual = CellUtil.cloneQualifier(cell); -2959 if (qual == null) qual = HConstants.EMPTY_BYTE_ARRAY; -2960 -2961 Integer count = kvCount.get(qual); -2962 if (count == null) { -2963kvCount.put(qual, 1); -2964 } else { -2965kvCount.put(qual, count + 1); -2966 } -2967 count = kvCount.get(qual); -2968 -2969 Get get = new Get(CellUtil.cloneRow(cell)); -2970 get.setMaxVersions(count); -2971 get.addColumn(family, qual); -2972 if (coprocessorHost != null) { -2973if (!coprocessorHost.prePrepareTimeStampForDeleteVersion(mutation, cell, -2974byteNow, get)) { -2975 updateDeleteLatestVersionTimeStamp(cell, get, count, byteNow); -2976} -2977 } else { -2978 updateDeleteLatestVersionTimeStamp(cell, get, count, byteNow); -2979 } -2980} else { -2981 CellUtil.updateLatestStamp(cell, byteNow, 0); -2982} -2983 } -2984} -2985 } -2986 -2987 void updateDeleteLatestVersionTimeStamp(Cell cell, Get get, int count, byte[] byteNow) -2988 throws IOException { -2989ListCell result = get(get, false); -2990 -2991if (result.size() count) { -2992 // Nothing to delete -2993 CellUtil.updateLatestStamp(cell, byteNow, 0); -2994 return; -2995} -2996if (result.size() count) { -2997 throw new RuntimeException("Unexpected size: " + result.size()); -2998} -2999Cell getCell = result.get(count - 1); -3000CellUtil.setTimestamp(cell, getCell.getTimestamp()); -3001 } -3002 -3003 @Override -3004 public void put(Put put) throws IOException { -3005checkReadOnly(); -3006 -3007// Do a rough check that we have resources to accept a write. The check is -3008// 'rough' in that between the resource check and the call to obtain a -3009// read lock, resources may run out. For now, the thought is that this -3010// will be extremely rare; we'll deal with it when it happens. -3011checkResources(); -3012 startRegionOperation(Operation.PUT); -3013try { -3014 // All edits for the given row (across all column families) must happen atomically. -3015 doBatchMutate(put); -3016} finally { -3017 closeRegionOperation(Operation.PUT); -3018} -3019 } -3020 -3021 /** -3022 * Struct-like class that tracks the progress of a batch operation, -3023 * accumulating status codes and tracking the index at which processing -3024 * is proceeding. -3025 */ -3026 private abstract static class BatchOperationT { -3027T[] operations; -3028int nextIndexToProcess = 0; -3029OperationStatus[] retCodeDetails; -3030WALEdit[] walEditsFromCoprocessors; -3031 -3032public BatchOperation(T[] operations) { -3033 this.operations = operations; -3034 this.retCodeDetails = new OperationStatus[operations.length]; -3035 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -3036 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -3037} -3038 -3039public abstract Mutation getMutation(int index); -3040public abstract long getNonceGroup(int index); -3041public abstract long getNonce(int index); -3042/** This method is potentially expensive and should only be used for non-replay CP path. */ -3043public abstract Mutation[] getMutationsForCoprocs(); -3044public abstract boolean isInReplay(); -3045public abstract long getReplaySequenceId(); -3046 -3047public boolean isDone() { -3048 return nextIndexToProcess == operations.length; -3049} -3050 } -3051 -3052 private static class MutationBatch extends BatchOperationMutation { -3053private long nonceGroup; -3054private long nonce; -3055public MutationBatch(Mutation[] operations, long nonceGroup, long nonce) { -3056 super(operations); -3057 this.nonceGroup = nonceGroup; -3058 this.nonce = nonce; -3059} -3060 -3061@Override -3062public Mutation getMutation(int index) { -3063 return this.operations[index]; -3064} -3065 -3066
[17/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MergeTableRegionProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MergeTableRegionProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MergeTableRegionProcedureBiConsumer.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MergeTableRegionProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MergeTableRegionProcedureBiConsumer.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName
[29/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html index 57963ea..8765ab3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/AsyncHBaseAdmin.html @@ -125,538 +125,513 @@ 117 } 118 119 @Override -120 public CompletableFutureListTableDescriptor deleteTables(Pattern pattern) { -121return wrap(rawAdmin.deleteTables(pattern)); +120 public CompletableFutureVoid truncateTable(TableName tableName, boolean preserveSplits) { +121return wrap(rawAdmin.truncateTable(tableName, preserveSplits)); 122 } 123 124 @Override -125 public CompletableFutureVoid truncateTable(TableName tableName, boolean preserveSplits) { -126return wrap(rawAdmin.truncateTable(tableName, preserveSplits)); +125 public CompletableFutureVoid enableTable(TableName tableName) { +126return wrap(rawAdmin.enableTable(tableName)); 127 } 128 129 @Override -130 public CompletableFutureVoid enableTable(TableName tableName) { -131return wrap(rawAdmin.enableTable(tableName)); +130 public CompletableFutureVoid disableTable(TableName tableName) { +131return wrap(rawAdmin.disableTable(tableName)); 132 } 133 134 @Override -135 public CompletableFutureListTableDescriptor enableTables(Pattern pattern) { -136return wrap(rawAdmin.enableTables(pattern)); +135 public CompletableFutureBoolean isTableEnabled(TableName tableName) { +136return wrap(rawAdmin.isTableEnabled(tableName)); 137 } 138 139 @Override -140 public CompletableFutureVoid disableTable(TableName tableName) { -141return wrap(rawAdmin.disableTable(tableName)); +140 public CompletableFutureBoolean isTableDisabled(TableName tableName) { +141return wrap(rawAdmin.isTableDisabled(tableName)); 142 } 143 144 @Override -145 public CompletableFutureListTableDescriptor disableTables(Pattern pattern) { -146return wrap(rawAdmin.disableTables(pattern)); +145 public CompletableFutureBoolean isTableAvailable(TableName tableName, byte[][] splitKeys) { +146return wrap(rawAdmin.isTableAvailable(tableName, splitKeys)); 147 } 148 149 @Override -150 public CompletableFutureBoolean isTableEnabled(TableName tableName) { -151return wrap(rawAdmin.isTableEnabled(tableName)); -152 } -153 -154 @Override -155 public CompletableFutureBoolean isTableDisabled(TableName tableName) { -156return wrap(rawAdmin.isTableDisabled(tableName)); -157 } -158 -159 @Override -160 public CompletableFutureBoolean isTableAvailable(TableName tableName, byte[][] splitKeys) { -161return wrap(rawAdmin.isTableAvailable(tableName, splitKeys)); -162 } -163 -164 @Override -165 public CompletableFuturePairInteger, Integer getAlterStatus(TableName tableName) { -166return wrap(rawAdmin.getAlterStatus(tableName)); -167 } -168 -169 @Override -170 public CompletableFutureVoid addColumnFamily(TableName tableName, -171 ColumnFamilyDescriptor columnFamily) { -172return wrap(rawAdmin.addColumnFamily(tableName, columnFamily)); -173 } -174 -175 @Override -176 public CompletableFutureVoid deleteColumnFamily(TableName tableName, byte[] columnFamily) { -177return wrap(rawAdmin.deleteColumnFamily(tableName, columnFamily)); -178 } -179 -180 @Override -181 public CompletableFutureVoid modifyColumnFamily(TableName tableName, -182 ColumnFamilyDescriptor columnFamily) { -183return wrap(rawAdmin.modifyColumnFamily(tableName, columnFamily)); +150 public CompletableFutureVoid addColumnFamily(TableName tableName, +151 ColumnFamilyDescriptor columnFamily) { +152return wrap(rawAdmin.addColumnFamily(tableName, columnFamily)); +153 } +154 +155 @Override +156 public CompletableFutureVoid deleteColumnFamily(TableName tableName, byte[] columnFamily) { +157return wrap(rawAdmin.deleteColumnFamily(tableName, columnFamily)); +158 } +159 +160 @Override +161 public CompletableFutureVoid modifyColumnFamily(TableName tableName, +162 ColumnFamilyDescriptor columnFamily) { +163return wrap(rawAdmin.modifyColumnFamily(tableName, columnFamily)); +164 } +165 +166 @Override +167 public CompletableFutureVoid createNamespace(NamespaceDescriptor descriptor) { +168return wrap(rawAdmin.createNamespace(descriptor)); +169 } +170 +171 @Override +172 public CompletableFutureVoid modifyNamespace(NamespaceDescriptor descriptor) { +173return wrap(rawAdmin.modifyNamespace(descriptor)); +174 } +175 +176 @Override +177 public CompletableFutureVoid deleteNamespace(String name) { +178return wrap(rawAdmin.deleteNamespace(name)); +179 } +180 +181 @Override +182 public
[34/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html index 640dd2a..32a39f8 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.BulkLoadListener.html @@ -109,7 +109,7 @@ var activeTableTab = "activeTableTab"; -public static interface HRegion.BulkLoadListener +public static interface HRegion.BulkLoadListener Listener class to enable callers of bulkLoadHFile() to perform any necessary pre/post processing of a given bulkload call @@ -174,7 +174,7 @@ var activeTableTab = "activeTableTab"; prepareBulkLoad -http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringprepareBulkLoad(byte[]family, +http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringprepareBulkLoad(byte[]family, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringsrcPath, booleancopyFile) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException @@ -196,7 +196,7 @@ var activeTableTab = "activeTableTab"; doneBulkLoad -voiddoneBulkLoad(byte[]family, +voiddoneBulkLoad(byte[]family, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringsrcPath) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Called after a successful HFile load @@ -215,7 +215,7 @@ var activeTableTab = "activeTableTab"; failedBulkLoad -voidfailedBulkLoad(byte[]family, +voidfailedBulkLoad(byte[]family, http://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringsrcPath) throws http://docs.oracle.com/javase/8/docs/api/java/io/IOException.html?is-external=true; title="class or interface in java.io">IOException Called after a failed HFile load http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html index eb1933e..bda20bb 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.MutationBatch.html @@ -118,7 +118,7 @@ var activeTableTab = "activeTableTab"; -private static class HRegion.MutationBatch +private static class HRegion.MutationBatch extends HRegion.BatchOperationMutation @@ -248,7 +248,7 @@ extends nonceGroup -privatelong nonceGroup +privatelong nonceGroup @@ -257,7 +257,7 @@ extends nonce -privatelong nonce +privatelong nonce @@ -274,7 +274,7 @@ extends MutationBatch -publicMutationBatch(Mutation[]operations, +publicMutationBatch(Mutation[]operations, longnonceGroup, longnonce) @@ -293,7 +293,7 @@ extends getMutation -publicMutationgetMutation(intindex) +publicMutationgetMutation(intindex) Specified by: getMutationin classHRegion.BatchOperationMutation @@ -306,7 +306,7 @@ extends getNonceGroup -publiclonggetNonceGroup(intindex) +publiclonggetNonceGroup(intindex) Specified by: getNonceGroupin classHRegion.BatchOperationMutation @@ -319,7 +319,7 @@ extends getNonce -publiclonggetNonce(intindex) +publiclonggetNonce(intindex) Specified by: getNoncein classHRegion.BatchOperationMutation @@ -332,7 +332,7 @@ extends getMutationsForCoprocs -publicMutation[]getMutationsForCoprocs() +publicMutation[]getMutationsForCoprocs() Description copied from class:HRegion.BatchOperation This method is potentially expensive and should only be used for non-replay CP path. @@ -347,7 +347,7 @@ extends isInReplay -publicbooleanisInReplay() +publicbooleanisInReplay() Specified by: isInReplayin classHRegion.BatchOperationMutation @@ -360,7 +360,7 @@ extends getReplaySequenceId -publiclonggetReplaySequenceId() +publiclonggetReplaySequenceId() Specified by: getReplaySequenceIdin classHRegion.BatchOperationMutation
[16/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyColumnFamilyProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyColumnFamilyProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyColumnFamilyProcedureBiConsumer.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyColumnFamilyProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyColumnFamilyProcedureBiConsumer.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor
[11/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.TableOperator.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName tableName) { -443 CompletableFutureTableDescriptor future = new CompletableFuture(); -444
[45/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/book.html -- diff --git a/book.html b/book.html index 2a4a579..ad11160 100644 --- a/book.html +++ b/book.html @@ -35225,7 +35225,7 @@ The server will return cellblocks compressed using this same compressor as long Version 3.0.0-SNAPSHOT -Last updated 2017-10-15 14:29:42 UTC +Last updated 2017-10-16 14:29:40 UTC http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/bulk-loads.html -- diff --git a/bulk-loads.html b/bulk-loads.html index 24eecdd..eef02c2 100644 --- a/bulk-loads.html +++ b/bulk-loads.html @@ -7,7 +7,7 @@ - + Apache HBase Bulk Loads in Apache HBase (TM) @@ -311,7 +311,7 @@ under the License. --> https://www.apache.org/;>The Apache Software Foundation. All rights reserved. - Last Published: 2017-10-15 + Last Published: 2017-10-16
[44/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/checkstyle-aggregate.html -- diff --git a/checkstyle-aggregate.html b/checkstyle-aggregate.html index 93e11d0..5a76ffe 100644 --- a/checkstyle-aggregate.html +++ b/checkstyle-aggregate.html @@ -7,7 +7,7 @@ - + Apache HBase Checkstyle Results @@ -289,7 +289,7 @@ 2055 0 0 -13608 +13609 Files @@ -887,7 +887,7 @@ org/apache/hadoop/hbase/client/AsyncAdmin.java 0 0 -18 +16 org/apache/hadoop/hbase/client/AsyncAdminBuilder.java 0 @@ -927,7 +927,7 @@ org/apache/hadoop/hbase/client/AsyncHBaseAdmin.java 0 0 -2 +3 org/apache/hadoop/hbase/client/AsyncMasterRequestRpcRetryingCaller.java 0 @@ -1277,7 +1277,7 @@ org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.java 0 0 -127 +130 org/apache/hadoop/hbase/client/RawAsyncTable.java 0 @@ -5032,7 +5032,7 @@ org/apache/hadoop/hbase/regionserver/HRegion.java 0 0 -228 +227 org/apache/hadoop/hbase/regionserver/HRegionFileSystem.java 0 @@ -8199,7 +8199,7 @@ http://checkstyle.sourceforge.net/config_blocks.html#NeedBraces;>NeedBraces -1692 +1691 Error coding @@ -8273,7 +8273,7 @@ http://checkstyle.sourceforge.net/config_imports.html#UnusedImports;>UnusedImports processJavadoc: true -117 +123 Error indentation @@ -8291,12 +8291,12 @@ http://checkstyle.sourceforge.net/config_javadoc.html#JavadocTagContinuationIndentation;>JavadocTagContinuationIndentation offset: 2 -763 +769 Error http://checkstyle.sourceforge.net/config_javadoc.html#NonEmptyAtclauseDescription;>NonEmptyAtclauseDescription -3271 +3265 Error misc @@ -8314,7 +8314,7 @@ max: 100 ignorePattern: ^package.*|^import.*|a href|href|http://|https://|ftp://|org.apache.thrift.|com.google.protobuf.|hbase.protobuf.generated -1102 +1098 Error @@ -14680,7 +14680,7 @@ Error javadoc -NonEmptyAtclauseDescription +JavadocTagContinuationIndentation Javadoc comment at column 43 has parse error. Missed HTML close tag 'TableName'. Sometimes it means that close tag missed for one of previous tags. 179 @@ -16645,111 +16645,99 @@ Error imports +UnusedImports +Unused import - org.apache.hadoop.hbase.util.Pair. +43 + +Error +imports ImportOrder Wrong order for 'com.google.protobuf.RpcChannel' import. 46 - -Error -sizes -LineLength -Line is longer than 100 characters (found 101). -171 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -181 - -Error -sizes -LineLength -Line is longer than 100 characters (found 101). -187 - -Error -sizes -LineLength -Line is longer than 100 characters (found 106). -190 +160 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -417 +363 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -431 +377 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -817 +763 Error sizes LineLength Line is longer than 100 characters (found 106). -854 +800 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -866 +812 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -885 +831 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -914 +860 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -921 +867 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -922 +868 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -979 +925 Error sizes LineLength Line is longer than 100 characters (found 101). -1006 +952 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -1034 +980 Error javadoc NonEmptyAtclauseDescription At-clause should have a non-empty description. -1055 +1001 org/apache/hadoop/hbase/client/AsyncAdminBuilder.java @@ -16981,31 +16969,37 @@ Error imports +UnusedImports +Unused import - org.apache.hadoop.hbase.util.Pair. +46 + +Error +imports ImportOrder Wrong order for 'com.google.protobuf.RpcChannel' import. 49 - + Error sizes LineLength Line is longer than 100 characters (found 101). -409 +384 org/apache/hadoop/hbase/client/AsyncMasterRequestRpcRetryingCaller.java - + Severity Category Rule Message Line - + Error imports ImportOrder Wrong order for 'java.util.concurrent.CompletableFuture' import. 22 - + Error imports ImportOrder @@ -17014,13 +17008,13 @@ org/apache/hadoop/hbase/client/AsyncMetaRegionLocator.java - + Severity Category Rule Message Line - + Error imports AvoidStarImport @@ -17029,49 +17023,49 @@ org/apache/hadoop/hbase/client/AsyncNonMetaRegionLocator.java - + Severity Category Rule Message Line - + Error imports
[33/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html -- diff --git a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html index deb8775..34a43f1 100644 --- a/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html +++ b/devapidocs/org/apache/hadoop/hbase/regionserver/HRegion.RowLockImpl.html @@ -117,7 +117,7 @@ var activeTableTab = "activeTableTab"; -public static class HRegion.RowLockImpl +public static class HRegion.RowLockImpl extends http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object implements Region.RowLock Class used to represent a lock on a row. @@ -226,7 +226,7 @@ implements context -private finalHRegion.RowLockContext context +private finalHRegion.RowLockContext context @@ -235,7 +235,7 @@ implements lock -private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Lock.html?is-external=true; title="class or interface in java.util.concurrent.locks">Lock lock +private finalhttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Lock.html?is-external=true; title="class or interface in java.util.concurrent.locks">Lock lock @@ -252,7 +252,7 @@ implements RowLockImpl -publicRowLockImpl(HRegion.RowLockContextcontext, +publicRowLockImpl(HRegion.RowLockContextcontext, http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Lock.html?is-external=true; title="class or interface in java.util.concurrent.locks">Locklock) @@ -270,7 +270,7 @@ implements getLock -publichttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Lock.html?is-external=true; title="class or interface in java.util.concurrent.locks">LockgetLock() +publichttp://docs.oracle.com/javase/8/docs/api/java/util/concurrent/locks/Lock.html?is-external=true; title="class or interface in java.util.concurrent.locks">LockgetLock() @@ -279,7 +279,7 @@ implements getContext -publicHRegion.RowLockContextgetContext() +publicHRegion.RowLockContextgetContext() @@ -288,7 +288,7 @@ implements release -publicvoidrelease() +publicvoidrelease() Description copied from interface:Region.RowLock Release the given lock. If there are no remaining locks held by the current thread then unlock the row and allow other threads to acquire the lock. @@ -304,7 +304,7 @@ implements toString -publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringtoString() +publichttp://docs.oracle.com/javase/8/docs/api/java/lang/String.html?is-external=true; title="class or interface in java.lang">StringtoString() Overrides: http://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true#toString--; title="class or interface in java.lang">toStringin classhttp://docs.oracle.com/javase/8/docs/api/java/lang/Object.html?is-external=true; title="class or interface in java.lang">Object
[22/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteNamespaceProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteNamespaceProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteNamespaceProcedureBiConsumer.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteNamespaceProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteNamespaceProcedureBiConsumer.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName
[50/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/apidocs/deprecated-list.html -- diff --git a/apidocs/deprecated-list.html b/apidocs/deprecated-list.html index d448000..7fbe5b7 100644 --- a/apidocs/deprecated-list.html +++ b/apidocs/deprecated-list.html @@ -327,182 +327,175 @@ -org.apache.hadoop.hbase.client.AsyncAdmin.closeRegion(byte[], OptionalServerName) -As of release 2.0.0, this will be removed in HBase 3.0.0 - (https://issues.apache.org/jira/browse/HBASE-18231;>HBASE-18231). - Use AsyncAdmin.unassign(byte[], boolean). - - - org.apache.hadoop.hbase.client.Admin.closeRegion(byte[], String) As of release 2.0.0, this will be removed in HBase 3.0.0. Use Admin.unassign(byte[], boolean). - + org.apache.hadoop.hbase.client.Admin.closeRegion(ServerName, HRegionInfo) As of release 2.0.0, this will be removed in HBase 3.0.0 (https://issues.apache.org/jira/browse/HBASE-18231;>HBASE-18231). Use Admin.unassign(byte[], boolean). - + org.apache.hadoop.hbase.client.Admin.closeRegion(String, String) As of release 2.0.0, this will be removed in HBase 3.0.0. Use Admin.unassign(byte[], boolean). - + org.apache.hadoop.hbase.client.Admin.closeRegionWithEncodedRegionName(String, String) As of release 2.0.0, this will be removed in HBase 3.0.0. Use Admin.unassign(byte[], boolean). - + org.apache.hadoop.hbase.filter.CompareFilter.compareFamily(CompareFilter.CompareOp, ByteArrayComparable, Cell) Since 2.0.0. Will be removed in 3.0.0. Use CompareFilter.compareFamily(CompareOperator, ByteArrayComparable, Cell) - + org.apache.hadoop.hbase.filter.CompareFilter.compareQualifier(CompareFilter.CompareOp, ByteArrayComparable, Cell) Since 2.0.0. Will be removed in 3.0.0. Use CompareFilter.compareQualifier(CompareOperator, ByteArrayComparable, Cell) - + org.apache.hadoop.hbase.filter.CompareFilter.compareRow(CompareFilter.CompareOp, ByteArrayComparable, Cell) Since 2.0.0. Will be removed in 3.0.0. Use CompareFilter.compareRow(CompareOperator, ByteArrayComparable, Cell) - + org.apache.hadoop.hbase.filter.CompareFilter.compareValue(CompareFilter.CompareOp, ByteArrayComparable, Cell) Since 2.0.0. Will be removed in 3.0.0. Use CompareFilter.compareValue(CompareOperator, ByteArrayComparable, Cell) - + org.apache.hadoop.hbase.HRegionInfo.convert(HBaseProtos.RegionInfo) As of release 2.0.0, this will be removed in HBase 3.0.0 Use toRegionInfo(HBaseProtos.RegionInfo) in org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil. - + org.apache.hadoop.hbase.HRegionInfo.convert(HRegionInfo) As of release 2.0.0, this will be removed in HBase 3.0.0 Use toRegionInfo(org.apache.hadoop.hbase.client.RegionInfo) in org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil. - + org.apache.hadoop.hbase.mapreduce.CellCreator.create(byte[], int, int, byte[], int, int, byte[], int, int, long, byte[], int, int, String) - + org.apache.hadoop.hbase.CellUtil.createCell(byte[]) As of release 2.0.0, this will be removed in HBase 3.0.0. Use CellBuilder instead - + org.apache.hadoop.hbase.CellUtil.createCell(byte[], byte[]) As of release 2.0.0, this will be removed in HBase 3.0.0. Use CellBuilder instead - + org.apache.hadoop.hbase.CellUtil.createCell(byte[], byte[], byte[]) As of release 2.0.0, this will be removed in HBase 3.0.0. Use CellBuilder instead - + org.apache.hadoop.hbase.CellUtil.createCell(byte[], byte[], byte[], long, byte, byte[]) As of release 2.0.0, this will be removed in HBase 3.0.0. Use CellBuilder instead - + org.apache.hadoop.hbase.CellUtil.createCell(byte[], byte[], byte[], long, byte, byte[], byte[], long) As of release 2.0.0, this will be removed in HBase 3.0.0. Use ExtendedCellBuilder instead - + org.apache.hadoop.hbase.CellUtil.createCell(byte[], byte[], byte[], long, byte, byte[], long) As of release 2.0.0, this will be removed in HBase 3.0.0. Use ExtendedCellBuilder instead - + org.apache.hadoop.hbase.CellUtil.createCell(byte[], byte[], byte[], long, KeyValue.Type, byte[], byte[]) As of release 2.0.0, this will be removed in HBase 3.0.0. Use ExtendedCellBuilder instead - + org.apache.hadoop.hbase.CellUtil.createCell(byte[], int, int, byte[], int, int, byte[], int, int) As of release 2.0.0, this will be removed in HBase 3.0.0. Use CellBuilder instead - + org.apache.hadoop.hbase.filter.ParseFilter.createCompareOp(byte[]) Since 2.0 - + org.apache.hadoop.hbase.filter.ParseFilter.createCompareOperator(byte[]) Since 2.0 - + org.apache.hadoop.hbase.HRegionInfo.createRegionName(TableName, byte[], byte[], boolean) As of
[20/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DisableTableProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DisableTableProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DisableTableProcedureBiConsumer.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DisableTableProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DisableTableProcedureBiConsumer.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName tableName) { -443
[48/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/apidocs/org/apache/hadoop/hbase/client/AsyncAdmin.html -- diff --git a/apidocs/org/apache/hadoop/hbase/client/AsyncAdmin.html b/apidocs/org/apache/hadoop/hbase/client/AsyncAdmin.html index 159a8a7..2a0edeb 100644 --- a/apidocs/org/apache/hadoop/hbase/client/AsyncAdmin.html +++ b/apidocs/org/apache/hadoop/hbase/client/AsyncAdmin.html @@ -18,8 +18,8 @@ catch(err) { } //--> -var methods = {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":18,"i6":6,"i7":6,"i8":6,"i9":6,"i10":38,"i11":18,"i12":6,"i13":18,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":18,"i20":6,"i21":6,"i22":6,"i23":6,"i24":6,"i25":18,"i26":6,"i27":6,"i28":6,"i29":6,"i30":6,"i31":6,"i32":6,"i33":6,"i34":6,"i35":6,"i36":6,"i37":6,"i38":6,"i39":6,"i40":6,"i41":18,"i42":6,"i43":6,"i44":6,"i45":6,"i46":6,"i47":6,"i48":6,"i49":18,"i50":6,"i51":6,"i52":6,"i53":6,"i54":18,"i55":6,"i56":18,"i57":6,"i58":6,"i59":6,"i60":6,"i61":6,"i62":6,"i63":6,"i64":6,"i65":6,"i66":6,"i67":6,"i68":6,"i69":6,"i70":18,"i71":6,"i72":6,"i73":6,"i74":6,"i75":6,"i76":6,"i77":6,"i78":18,"i79":6,"i80":18,"i81":6,"i82":18,"i83":6,"i84":18,"i85":6,"i86":6,"i87":18,"i88":6,"i89":18,"i90":6,"i91":6,"i92":6,"i93":6,"i94":6,"i95":6,"i96":6,"i97":6,"i98":6,"i99":6,"i100":6,"i101":6,"i102":6,"i103":6,"i104":6,"i105":6,"i106":6,"i107":6,"i108":6,"i109":6,"i110":6,"i111":6,"i112":6,"i113":6,"i114":6,"i115":18,"i116":18,"i117":6,"i118":6,"i 119":18,"i120":6,"i121":6,"i122":6,"i123":6,"i124":6,"i125":6,"i126":6,"i127":6,"i128":6}; -var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],16:["t5","Default Methods"],32:["t6","Deprecated Methods"]}; +var methods = {"i0":6,"i1":6,"i2":6,"i3":6,"i4":6,"i5":18,"i6":6,"i7":6,"i8":6,"i9":6,"i10":18,"i11":6,"i12":18,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":18,"i19":6,"i20":6,"i21":6,"i22":6,"i23":6,"i24":18,"i25":6,"i26":6,"i27":6,"i28":6,"i29":6,"i30":6,"i31":6,"i32":6,"i33":6,"i34":6,"i35":6,"i36":18,"i37":6,"i38":6,"i39":6,"i40":6,"i41":6,"i42":6,"i43":6,"i44":18,"i45":6,"i46":6,"i47":6,"i48":6,"i49":18,"i50":6,"i51":18,"i52":6,"i53":6,"i54":6,"i55":6,"i56":6,"i57":6,"i58":6,"i59":6,"i60":6,"i61":6,"i62":6,"i63":6,"i64":6,"i65":18,"i66":6,"i67":6,"i68":6,"i69":6,"i70":6,"i71":6,"i72":6,"i73":18,"i74":6,"i75":18,"i76":6,"i77":18,"i78":6,"i79":18,"i80":6,"i81":6,"i82":18,"i83":6,"i84":18,"i85":6,"i86":6,"i87":6,"i88":6,"i89":6,"i90":6,"i91":6,"i92":6,"i93":6,"i94":6,"i95":6,"i96":6,"i97":6,"i98":6,"i99":6,"i100":6,"i101":6,"i102":6,"i103":6,"i104":6,"i105":6,"i106":6,"i107":6,"i108":6,"i109":6,"i110":18,"i111":18,"i112":6,"i113":6,"i114":18,"i115":6,"i116":6,"i117":6,"i118":6,"i 119":6,"i120":6,"i121":6,"i122":6,"i123":6}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],16:["t5","Default Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; var tableTab = "tableTab"; @@ -124,7 +124,7 @@ public interface -All MethodsInstance MethodsAbstract MethodsDefault MethodsDeprecated Methods +All MethodsInstance MethodsAbstract MethodsDefault Methods Modifier and Type Method and Description @@ -194,56 +194,45 @@ public interface -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean -closeRegion(byte[]regionName, - http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">OptionalServerNameserverName) -Deprecated. -As of release 2.0.0, this will be removed in HBase 3.0.0 - (https://issues.apache.org/jira/browse/HBASE-18231;>HBASE-18231). - Use unassign(byte[], boolean). - - - - default http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void compact(TableNametableName) Compact a table. - + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void compact(TableNametableName, http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">Optionalbyte[]columnFamily) Compact a column family within a table. - + default
[28/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.AddColumnFamilyProcedureBiConsumer.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName
[15/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.ModifyNamespaceProcedureBiConsumer.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName
[49/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/apidocs/index-all.html -- diff --git a/apidocs/index-all.html b/apidocs/index-all.html index aee98f8..652f63a 100644 --- a/apidocs/index-all.html +++ b/apidocs/index-all.html @@ -1509,14 +1509,6 @@ Use Admin.unassign(byte[], boolean). -closeRegion(byte[], OptionalServerName) - Method in interface org.apache.hadoop.hbase.client.AsyncAdmin - -Deprecated. -As of release 2.0.0, this will be removed in HBase 3.0.0 - (https://issues.apache.org/jira/browse/HBASE-18231;>HBASE-18231). - Use AsyncAdmin.unassign(byte[], boolean). - - closeRegionWithEncodedRegionName(String, String) - Method in interface org.apache.hadoop.hbase.client.Admin Deprecated. @@ -3899,10 +3891,6 @@ and Admin.deleteTable(TableName) -deleteTables(Pattern) - Method in interface org.apache.hadoop.hbase.client.AsyncAdmin - -Delete tables matching the passed in pattern and wait on completion. - deleteTableSnapshots(String, String) - Method in interface org.apache.hadoop.hbase.client.Admin Deprecated. @@ -4049,10 +4037,6 @@ and Admin.disableTable(org.apache.hadoop.hbase.TableName) -disableTables(Pattern) - Method in interface org.apache.hadoop.hbase.client.AsyncAdmin - -Disable tables matching the passed in pattern. - DISALLOW_WRITES_IN_RECOVERING - Static variable in class org.apache.hadoop.hbase.HConstants DISALLOWED_TABLE_NAME - Static variable in class org.apache.hadoop.hbase.TableName @@ -4281,10 +4265,6 @@ and Admin.enableTable(org.apache.hadoop.hbase.TableName) -enableTables(Pattern) - Method in interface org.apache.hadoop.hbase.client.AsyncAdmin - -Enable tables matching the passed in pattern. - ENC_SEPARATOR - Static variable in interface org.apache.hadoop.hbase.client.RegionInfo Separator used to demarcate the encodedName in a region name @@ -5609,11 +5589,6 @@ on an operation. -getAlterStatus(TableName) - Method in interface org.apache.hadoop.hbase.client.AsyncAdmin - -Get the status of alter command - indicates how many regions have received the updated schema - Asynchronous operation. - getAttribute(String) - Method in interface org.apache.hadoop.hbase.client.Attributes Gets an attribute http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html -- diff --git a/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html b/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html index d24a59b..15f45b8 100644 --- a/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html +++ b/apidocs/org/apache/hadoop/hbase/class-use/ServerName.html @@ -548,42 +548,31 @@ -http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Boolean.html?is-external=true; title="class or interface in java.lang">Boolean -AsyncAdmin.closeRegion(byte[]regionName, - http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">OptionalServerNameserverName) -Deprecated. -As of release 2.0.0, this will be removed in HBase 3.0.0 - (https://issues.apache.org/jira/browse/HBASE-18231;>HBASE-18231). - Use AsyncAdmin.unassign(byte[], boolean). - - - - http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void AsyncAdmin.drainRegionServers(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerNameservers) Mark a region server as draining to prevent additional regions from getting assigned to it. - + void Admin.drainRegionServers(http://docs.oracle.com/javase/8/docs/api/java/util/List.html?is-external=true; title="class or interface in java.util">ListServerNameservers) Mark a region server as draining to prevent additional regions from getting assigned to it. - + http://docs.oracle.com/javase/8/docs/api/java/util/concurrent/CompletableFuture.html?is-external=true; title="class or interface in java.util.concurrent">CompletableFuturehttp://docs.oracle.com/javase/8/docs/api/java/lang/Void.html?is-external=true; title="class or interface in java.lang">Void AsyncAdmin.move(byte[]regionName, http://docs.oracle.com/javase/8/docs/api/java/util/Optional.html?is-external=true; title="class or interface in java.util">OptionalServerNamedestServerName) Move the region r to dest. - +
[14/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.NamespaceProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.NamespaceProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.NamespaceProcedureBiConsumer.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.NamespaceProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.NamespaceProcedureBiConsumer.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName tableName) { -443
[08/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName tableName) { -443 CompletableFutureTableDescriptor future = new CompletableFuture(); -444this.ListTableSchema newMasterCaller() -445.action( -446
[05/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html index 81d256e..73c1480 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/regionserver/HRegion.FlushResult.Result.html @@ -2964,5322 +2964,5321 @@ 2956// This is expensive. 2957if (cell.getTimestamp() == HConstants.LATEST_TIMESTAMP CellUtil.isDeleteType(cell)) { 2958 byte[] qual = CellUtil.cloneQualifier(cell); -2959 if (qual == null) qual = HConstants.EMPTY_BYTE_ARRAY; -2960 -2961 Integer count = kvCount.get(qual); -2962 if (count == null) { -2963kvCount.put(qual, 1); -2964 } else { -2965kvCount.put(qual, count + 1); -2966 } -2967 count = kvCount.get(qual); -2968 -2969 Get get = new Get(CellUtil.cloneRow(cell)); -2970 get.setMaxVersions(count); -2971 get.addColumn(family, qual); -2972 if (coprocessorHost != null) { -2973if (!coprocessorHost.prePrepareTimeStampForDeleteVersion(mutation, cell, -2974byteNow, get)) { -2975 updateDeleteLatestVersionTimeStamp(cell, get, count, byteNow); -2976} -2977 } else { -2978 updateDeleteLatestVersionTimeStamp(cell, get, count, byteNow); -2979 } -2980} else { -2981 CellUtil.updateLatestStamp(cell, byteNow, 0); -2982} -2983 } -2984} -2985 } -2986 -2987 void updateDeleteLatestVersionTimeStamp(Cell cell, Get get, int count, byte[] byteNow) -2988 throws IOException { -2989ListCell result = get(get, false); -2990 -2991if (result.size() count) { -2992 // Nothing to delete -2993 CellUtil.updateLatestStamp(cell, byteNow, 0); -2994 return; -2995} -2996if (result.size() count) { -2997 throw new RuntimeException("Unexpected size: " + result.size()); -2998} -2999Cell getCell = result.get(count - 1); -3000CellUtil.setTimestamp(cell, getCell.getTimestamp()); -3001 } -3002 -3003 @Override -3004 public void put(Put put) throws IOException { -3005checkReadOnly(); -3006 -3007// Do a rough check that we have resources to accept a write. The check is -3008// 'rough' in that between the resource check and the call to obtain a -3009// read lock, resources may run out. For now, the thought is that this -3010// will be extremely rare; we'll deal with it when it happens. -3011checkResources(); -3012 startRegionOperation(Operation.PUT); -3013try { -3014 // All edits for the given row (across all column families) must happen atomically. -3015 doBatchMutate(put); -3016} finally { -3017 closeRegionOperation(Operation.PUT); -3018} -3019 } -3020 -3021 /** -3022 * Struct-like class that tracks the progress of a batch operation, -3023 * accumulating status codes and tracking the index at which processing -3024 * is proceeding. -3025 */ -3026 private abstract static class BatchOperationT { -3027T[] operations; -3028int nextIndexToProcess = 0; -3029OperationStatus[] retCodeDetails; -3030WALEdit[] walEditsFromCoprocessors; -3031 -3032public BatchOperation(T[] operations) { -3033 this.operations = operations; -3034 this.retCodeDetails = new OperationStatus[operations.length]; -3035 this.walEditsFromCoprocessors = new WALEdit[operations.length]; -3036 Arrays.fill(this.retCodeDetails, OperationStatus.NOT_RUN); -3037} -3038 -3039public abstract Mutation getMutation(int index); -3040public abstract long getNonceGroup(int index); -3041public abstract long getNonce(int index); -3042/** This method is potentially expensive and should only be used for non-replay CP path. */ -3043public abstract Mutation[] getMutationsForCoprocs(); -3044public abstract boolean isInReplay(); -3045public abstract long getReplaySequenceId(); -3046 -3047public boolean isDone() { -3048 return nextIndexToProcess == operations.length; -3049} -3050 } -3051 -3052 private static class MutationBatch extends BatchOperationMutation { -3053private long nonceGroup; -3054private long nonce; -3055public MutationBatch(Mutation[] operations, long nonceGroup, long nonce) { -3056 super(operations); -3057 this.nonceGroup = nonceGroup; -3058 this.nonce = nonce; -3059} -3060 -3061@Override -3062public Mutation getMutation(int index) { -3063 return this.operations[index]; -3064} -3065
[23/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteColumnFamilyProcedureBiConsumer.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteColumnFamilyProcedureBiConsumer.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteColumnFamilyProcedureBiConsumer.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteColumnFamilyProcedureBiConsumer.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.DeleteColumnFamilyProcedureBiConsumer.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor
[18/51] [partial] hbase-site git commit: Published site at .
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/47abd8e6/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MasterRpcCall.html -- diff --git a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MasterRpcCall.html b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MasterRpcCall.html index 6150454..b712da3 100644 --- a/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MasterRpcCall.html +++ b/devapidocs/src-html/org/apache/hadoop/hbase/client/RawAsyncHBaseAdmin.MasterRpcCall.html @@ -391,2559 +391,2494 @@ 383CompletableFutureVoid operate(TableName table); 384 } 385 -386 private CompletableFutureListTableDescriptor batchTableOperations(Pattern pattern, -387 TableOperator operator, String operationType) { -388 CompletableFutureListTableDescriptor future = new CompletableFuture(); -389ListTableDescriptor failed = new LinkedList(); -390 listTables(Optional.ofNullable(pattern), false).whenComplete( -391 (tables, error) - { -392if (error != null) { -393 future.completeExceptionally(error); -394 return; -395} -396CompletableFuture[] futures = -397tables.stream() -398.map((table) - operator.operate(table.getTableName()).whenComplete((v, ex) - { -399 if (ex != null) { -400LOG.info("Failed to " + operationType + " table " + table.getTableName(), ex); -401failed.add(table); -402 } -403 })).CompletableFuture toArray(size - new CompletableFuture[size]); -404 CompletableFuture.allOf(futures).thenAccept((v) - { -405 future.complete(failed); -406}); -407 }); -408return future; -409 } -410 -411 @Override -412 public CompletableFutureBoolean tableExists(TableName tableName) { -413return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +386 @Override +387 public CompletableFutureBoolean tableExists(TableName tableName) { +388return AsyncMetaTableAccessor.tableExists(metaTable, tableName); +389 } +390 +391 @Override +392 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, +393 boolean includeSysTables) { +394return this.ListTableDescriptor newMasterCaller() +395.action((controller, stub) - this +396 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( +397 controller, stub, +398 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), +399 (s, c, req, done) - s.getTableDescriptors(c, req, done), +400 (resp) - ProtobufUtil.toTableDescriptorList(resp))) +401.call(); +402 } +403 +404 @Override +405 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, +406 boolean includeSysTables) { +407return this.ListTableName newMasterCaller() +408.action((controller, stub) - this +409.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, +410 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), +411 (s, c, req, done) - s.getTableNames(c, req, done), +412 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( +413.call(); 414 } 415 416 @Override -417 public CompletableFutureListTableDescriptor listTables(OptionalPattern pattern, -418 boolean includeSysTables) { -419return this.ListTableDescriptor newMasterCaller() -420.action((controller, stub) - this -421 .GetTableDescriptorsRequest, GetTableDescriptorsResponse, ListTableDescriptor call( -422 controller, stub, -423 RequestConverter.buildGetTableDescriptorsRequest(pattern, includeSysTables), -424 (s, c, req, done) - s.getTableDescriptors(c, req, done), -425 (resp) - ProtobufUtil.toTableDescriptorList(resp))) -426.call(); -427 } -428 -429 @Override -430 public CompletableFutureListTableName listTableNames(OptionalPattern pattern, -431 boolean includeSysTables) { -432return this.ListTableName newMasterCaller() -433.action((controller, stub) - this -434.GetTableNamesRequest, GetTableNamesResponse, ListTableName call(controller, stub, -435 RequestConverter.buildGetTableNamesRequest(pattern, includeSysTables), -436 (s, c, req, done) - s.getTableNames(c, req, done), -437 (resp) - ProtobufUtil.toTableNameList(resp.getTableNamesList( -438.call(); -439 } -440 -441 @Override -442 public CompletableFutureTableDescriptor getTableDescriptor(TableName tableName) { -443 CompletableFutureTableDescriptor future = new CompletableFuture(); -444
hbase git commit: HBASE-18914 Remove AsyncAdmin's methods which were already deprecated in Admin interface
Repository: hbase Updated Branches: refs/heads/branch-2 aeaf222e3 -> 58b0585d6 HBASE-18914 Remove AsyncAdmin's methods which were already deprecated in Admin interface Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/58b0585d Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/58b0585d Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/58b0585d Branch: refs/heads/branch-2 Commit: 58b0585d66b90d3cdf47da84c64f2912e1773934 Parents: aeaf222 Author: Guanghao ZhangAuthored: Mon Oct 16 17:12:37 2017 +0800 Committer: Guanghao Zhang Committed: Mon Oct 16 22:23:46 2017 +0800 -- .../apache/hadoop/hbase/client/AsyncAdmin.java | 54 --- .../hadoop/hbase/client/AsyncHBaseAdmin.java| 25 --- .../hadoop/hbase/client/RawAsyncHBaseAdmin.java | 65 -- .../hbase/client/TestAsyncRegionAdminApi.java | 70 .../hbase/client/TestAsyncSnapshotAdminApi.java | 11 ++- .../hbase/client/TestAsyncTableAdminApi.java| 25 ++- 6 files changed, 9 insertions(+), 241 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/58b0585d/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncAdmin.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncAdmin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncAdmin.java index e539c6a..0a88138 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncAdmin.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncAdmin.java @@ -143,17 +143,6 @@ public interface AsyncAdmin { CompletableFuture deleteTable(TableName tableName); /** - * Delete tables matching the passed in pattern and wait on completion. Warning: Use this method - * carefully, there is no prompting and the effect is immediate. Consider using - * {@link #listTableNames(Optional, boolean) } and - * {@link #deleteTable(org.apache.hadoop.hbase.TableName)} - * @param pattern The pattern to match table names against - * @return Table descriptors for tables that couldn't be deleted. The return value will be wrapped - * by a {@link CompletableFuture}. The return HTDs are read-only. - */ - CompletableFuture deleteTables(Pattern pattern); - - /** * Truncate a table. * @param tableName name of table to truncate * @param preserveSplits True if the splits should be preserved @@ -167,32 +156,12 @@ public interface AsyncAdmin { CompletableFuture enableTable(TableName tableName); /** - * Enable tables matching the passed in pattern. Warning: Use this method carefully, there is no - * prompting and the effect is immediate. Consider using {@link #listTables(Optional, boolean)} and - * {@link #enableTable(TableName)} - * @param pattern The pattern to match table names against - * @return Table descriptors for tables that couldn't be enabled. The return value will be wrapped - * by a {@link CompletableFuture}. The return HTDs are read-only. - */ - CompletableFuture
enableTables(Pattern pattern); - - /** * Disable a table. The table has to be in enabled state for it to be disabled. * @param tableName */ CompletableFuture disableTable(TableName tableName); /** - * Disable tables matching the passed in pattern. Warning: Use this method carefully, there is no - * prompting and the effect is immediate. Consider using {@link #listTables(Optional, boolean)} and - * {@link #disableTable(TableName)} - * @param pattern The pattern to match table names against - * @return Table descriptors for tables that couldn't be disabled. The return value will be wrapped by a - * {@link CompletableFuture}. The return HTDs are read-only. - */ - CompletableFuture
disableTables(Pattern pattern); - - /** * @param tableName name of table to check * @return true if table is on-line. The return value will be wrapped by a * {@link CompletableFuture}. @@ -226,16 +195,6 @@ public interface AsyncAdmin { CompletableFuture isTableAvailable(TableName tableName, byte[][] splitKeys); /** - * Get the status of alter command - indicates how many regions have received the updated schema - * Asynchronous operation. - * @param tableName TableName instance - * @return Pair indicating the number of regions updated Pair.getFirst() is the regions that are - * yet to be updated Pair.getSecond() is the total number of regions of the table. The - * return value will be wrapped by a {@link CompletableFuture}. - */ - CompletableFuture
> getAlterStatus(TableName tableName); - - /**
hbase git commit: HBASE-18914 Remove AsyncAdmin's methods which were already deprecated in Admin interface
Repository: hbase Updated Branches: refs/heads/master 83af5f2c6 -> 51489b208 HBASE-18914 Remove AsyncAdmin's methods which were already deprecated in Admin interface Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/51489b20 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/51489b20 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/51489b20 Branch: refs/heads/master Commit: 51489b2081102b02785c89b6a03d36f54e29657b Parents: 83af5f2 Author: Guanghao ZhangAuthored: Mon Oct 16 17:12:37 2017 +0800 Committer: Guanghao Zhang Committed: Mon Oct 16 22:16:04 2017 +0800 -- .../apache/hadoop/hbase/client/AsyncAdmin.java | 54 --- .../hadoop/hbase/client/AsyncHBaseAdmin.java| 25 --- .../hadoop/hbase/client/RawAsyncHBaseAdmin.java | 65 -- .../hbase/client/TestAsyncRegionAdminApi.java | 70 .../hbase/client/TestAsyncSnapshotAdminApi.java | 11 ++- .../hbase/client/TestAsyncTableAdminApi.java| 25 ++- 6 files changed, 9 insertions(+), 241 deletions(-) -- http://git-wip-us.apache.org/repos/asf/hbase/blob/51489b20/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncAdmin.java -- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncAdmin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncAdmin.java index e539c6a..0a88138 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncAdmin.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncAdmin.java @@ -143,17 +143,6 @@ public interface AsyncAdmin { CompletableFuture deleteTable(TableName tableName); /** - * Delete tables matching the passed in pattern and wait on completion. Warning: Use this method - * carefully, there is no prompting and the effect is immediate. Consider using - * {@link #listTableNames(Optional, boolean) } and - * {@link #deleteTable(org.apache.hadoop.hbase.TableName)} - * @param pattern The pattern to match table names against - * @return Table descriptors for tables that couldn't be deleted. The return value will be wrapped - * by a {@link CompletableFuture}. The return HTDs are read-only. - */ - CompletableFuture deleteTables(Pattern pattern); - - /** * Truncate a table. * @param tableName name of table to truncate * @param preserveSplits True if the splits should be preserved @@ -167,32 +156,12 @@ public interface AsyncAdmin { CompletableFuture enableTable(TableName tableName); /** - * Enable tables matching the passed in pattern. Warning: Use this method carefully, there is no - * prompting and the effect is immediate. Consider using {@link #listTables(Optional, boolean)} and - * {@link #enableTable(TableName)} - * @param pattern The pattern to match table names against - * @return Table descriptors for tables that couldn't be enabled. The return value will be wrapped - * by a {@link CompletableFuture}. The return HTDs are read-only. - */ - CompletableFuture
enableTables(Pattern pattern); - - /** * Disable a table. The table has to be in enabled state for it to be disabled. * @param tableName */ CompletableFuture disableTable(TableName tableName); /** - * Disable tables matching the passed in pattern. Warning: Use this method carefully, there is no - * prompting and the effect is immediate. Consider using {@link #listTables(Optional, boolean)} and - * {@link #disableTable(TableName)} - * @param pattern The pattern to match table names against - * @return Table descriptors for tables that couldn't be disabled. The return value will be wrapped by a - * {@link CompletableFuture}. The return HTDs are read-only. - */ - CompletableFuture
disableTables(Pattern pattern); - - /** * @param tableName name of table to check * @return true if table is on-line. The return value will be wrapped by a * {@link CompletableFuture}. @@ -226,16 +195,6 @@ public interface AsyncAdmin { CompletableFuture isTableAvailable(TableName tableName, byte[][] splitKeys); /** - * Get the status of alter command - indicates how many regions have received the updated schema - * Asynchronous operation. - * @param tableName TableName instance - * @return Pair indicating the number of regions updated Pair.getFirst() is the regions that are - * yet to be updated Pair.getSecond() is the total number of regions of the table. The - * return value will be wrapped by a {@link CompletableFuture}. - */ - CompletableFuture
> getAlterStatus(TableName tableName); - - /** *