HBASE-17336 get/update replication peer config requests should be routed through master
Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/0e486656 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/0e486656 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/0e486656 Branch: refs/heads/hbase-12439 Commit: 0e48665641b16cd9b250503696b926a568063654 Parents: 1c477b2 Author: Guanghao Zhang <[email protected]> Authored: Tue Dec 27 14:20:57 2016 +0800 Committer: Guanghao Zhang <[email protected]> Committed: Fri Dec 30 10:12:47 2016 +0800 ---------------------------------------------------------------------- .../hbase/ReplicationPeerNotFoundException.java | 36 + .../org/apache/hadoop/hbase/client/Admin.java | 20 + .../hbase/client/ConnectionImplementation.java | 17 + .../apache/hadoop/hbase/client/HBaseAdmin.java | 28 + .../client/replication/ReplicationAdmin.java | 48 +- .../hbase/shaded/protobuf/RequestConverter.java | 18 + .../shaded/protobuf/generated/MasterProtos.java | 290 +- .../protobuf/generated/ReplicationProtos.java | 2509 +++++++++++++++++- .../src/main/protobuf/Master.proto | 8 + .../src/main/protobuf/Replication.proto | 17 + .../hbase/coprocessor/MasterObserver.java | 42 + .../org/apache/hadoop/hbase/master/HMaster.java | 27 + .../hbase/master/MasterCoprocessorHost.java | 42 + .../hadoop/hbase/master/MasterRpcServices.java | 33 + .../hadoop/hbase/master/MasterServices.java | 16 + .../master/replication/ReplicationManager.java | 23 +- .../hbase/security/access/AccessController.java | 13 + .../hbase/util/ServerRegionReplicaUtil.java | 19 +- .../replication/TestReplicationAdmin.java | 5 +- .../hbase/master/MockNoopMasterServices.java | 11 + .../TestRegionReplicaReplicationEndpoint.java | 35 +- .../security/access/TestAccessController.java | 31 + .../src/main/ruby/hbase/replication_admin.rb | 2 +- .../asciidoc/_chapters/appendix_acl_matrix.adoc | 2 + 24 files changed, 3194 insertions(+), 98 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hbase/blob/0e486656/hbase-client/src/main/java/org/apache/hadoop/hbase/ReplicationPeerNotFoundException.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ReplicationPeerNotFoundException.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ReplicationPeerNotFoundException.java new file mode 100644 index 0000000..daf7dd5 --- /dev/null +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ReplicationPeerNotFoundException.java @@ -0,0 +1,36 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.hadoop.hbase; + +import org.apache.hadoop.hbase.classification.InterfaceAudience; +import org.apache.hadoop.hbase.classification.InterfaceStability; + +/** + * Thrown when a replication peer can not be found + */ [email protected] [email protected] +public class ReplicationPeerNotFoundException extends DoNotRetryIOException { + + private static final long serialVersionUID = 1L; + + public ReplicationPeerNotFoundException(String peerId) { + super(peerId); + } +} \ No newline at end of file http://git-wip-us.apache.org/repos/asf/hbase/blob/0e486656/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java index fe3960f..2a1b782 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java @@ -1860,6 +1860,26 @@ public interface Admin extends Abortable, Closeable { } /** + * Returns the configured ReplicationPeerConfig for the specified peer + * @param peerId a short name that identifies the peer + * @return ReplicationPeerConfig for the peer + * @throws IOException + */ + default ReplicationPeerConfig getReplicationPeerConfig(final String peerId) throws IOException { + return new ReplicationPeerConfig(); + } + + /** + * Update the peerConfig for the specified peer + * @param peerId a short name that identifies the peer + * @param peerConfig new config for the peer + * @throws IOException + */ + default void updateReplicationPeerConfig(final String peerId, + final ReplicationPeerConfig peerConfig) throws IOException { + } + + /** * Mark a region server as draining to prevent additional regions from getting assigned to it. * @param servers List of region servers to drain. */ http://git-wip-us.apache.org/repos/asf/hbase/blob/0e486656/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionImplementation.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionImplementation.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionImplementation.java index ceac3fb..ea11c25 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionImplementation.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionImplementation.java @@ -100,8 +100,12 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.Disab import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse; import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest; import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse; +import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest; +import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse; import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest; import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse; +import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest; +import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; import org.apache.hadoop.hbase.util.ExceptionUtil; @@ -1694,6 +1698,19 @@ class ConnectionImplementation implements ClusterConnection, Closeable { throws ServiceException { return stub.removeDrainFromRegionServers(controller, request); } + + @Override + public GetReplicationPeerConfigResponse getReplicationPeerConfig(RpcController controller, + GetReplicationPeerConfigRequest request) throws ServiceException { + return stub.getReplicationPeerConfig(controller, request); + } + + @Override + public UpdateReplicationPeerConfigResponse updateReplicationPeerConfig( + RpcController controller, UpdateReplicationPeerConfigRequest request) + throws ServiceException { + return stub.updateReplicationPeerConfig(controller, request); + } }; } http://git-wip-us.apache.org/repos/asf/hbase/blob/0e486656/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java index ec4a5c1..89d1b49 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java @@ -67,6 +67,7 @@ import org.apache.hadoop.hbase.UnknownRegionException; import org.apache.hadoop.hbase.ZooKeeperConnectionException; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.classification.InterfaceStability; +import org.apache.hadoop.hbase.client.replication.ReplicationSerDeHelper; import org.apache.hadoop.hbase.client.security.SecurityCapability; import org.apache.hadoop.hbase.exceptions.TimeoutIOException; import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel; @@ -171,6 +172,7 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTa import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse; import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest; import org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos; +import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse; import org.apache.hadoop.hbase.snapshot.ClientSnapshotDescriptionUtils; import org.apache.hadoop.hbase.snapshot.HBaseSnapshotException; import org.apache.hadoop.hbase.snapshot.RestoreSnapshotException; @@ -3799,6 +3801,32 @@ public class HBaseAdmin implements Admin { } @Override + public ReplicationPeerConfig getReplicationPeerConfig(final String peerId) throws IOException { + return executeCallable(new MasterCallable<ReplicationPeerConfig>(getConnection(), + getRpcControllerFactory()) { + @Override + protected ReplicationPeerConfig rpcCall() throws Exception { + GetReplicationPeerConfigResponse response = master.getReplicationPeerConfig( + getRpcController(), RequestConverter.buildGetReplicationPeerConfigRequest(peerId)); + return ReplicationSerDeHelper.convert(response.getPeerConfig()); + } + }); + } + + @Override + public void updateReplicationPeerConfig(final String peerId, + final ReplicationPeerConfig peerConfig) throws IOException { + executeCallable(new MasterCallable<Void>(getConnection(), getRpcControllerFactory()) { + @Override + protected Void rpcCall() throws Exception { + master.updateReplicationPeerConfig(getRpcController(), + RequestConverter.buildUpdateReplicationPeerConfigRequest(peerId, peerConfig)); + return null; + } + }); + } + + @Override public void drainRegionServers(List<ServerName> servers) throws IOException { final List<HBaseProtos.ServerName> pbServers = new ArrayList<HBaseProtos.ServerName>(); for (ServerName server : servers) { http://git-wip-us.apache.org/repos/asf/hbase/blob/0e486656/hbase-client/src/main/java/org/apache/hadoop/hbase/client/replication/ReplicationAdmin.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/replication/ReplicationAdmin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/replication/ReplicationAdmin.java index 2d6c37b..d0859a4 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/replication/ReplicationAdmin.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/replication/ReplicationAdmin.java @@ -207,11 +207,8 @@ public class ReplicationAdmin implements Closeable { return ReplicationSerDeHelper.parseTableCFsFromConfig(tableCFsConfig); } - public void updatePeerConfig(String id, ReplicationPeerConfig peerConfig) - throws ReplicationException { - checkNamespacesAndTableCfsConfigConflict(peerConfig.getNamespaces(), - peerConfig.getTableCFsMap()); - this.replicationPeers.updatePeerConfig(id, peerConfig); + public void updatePeerConfig(String id, ReplicationPeerConfig peerConfig) throws IOException { + this.admin.updateReplicationPeerConfig(id, peerConfig); } /** @@ -250,8 +247,8 @@ public class ReplicationAdmin implements Closeable { return this.replicationPeers.getAllPeerConfigs(); } - public ReplicationPeerConfig getPeerConfig(String id) throws ReplicationException { - return this.replicationPeers.getReplicationPeerConfig(id); + public ReplicationPeerConfig getPeerConfig(String id) throws IOException { + return admin.getReplicationPeerConfig(id); } /** @@ -261,8 +258,9 @@ public class ReplicationAdmin implements Closeable { * use {@link #getPeerConfig(String)} instead. * */ @Deprecated - public String getPeerTableCFs(String id) throws ReplicationException { - return ReplicationSerDeHelper.convertToString(this.replicationPeers.getPeerTableCFsConfig(id)); + public String getPeerTableCFs(String id) throws IOException { + ReplicationPeerConfig peerConfig = admin.getReplicationPeerConfig(id); + return ReplicationSerDeHelper.convertToString(peerConfig.getTableCFsMap()); } /** @@ -270,11 +268,13 @@ public class ReplicationAdmin implements Closeable { * @param id a short that identifies the cluster * @param tableCfs table-cfs config str * @throws ReplicationException + * @throws IOException * @deprecated as release of 2.0.0, and it will be removed in 3.0.0, * use {@link #appendPeerTableCFs(String, Map)} instead. */ @Deprecated - public void appendPeerTableCFs(String id, String tableCfs) throws ReplicationException { + public void appendPeerTableCFs(String id, String tableCfs) throws ReplicationException, + IOException { appendPeerTableCFs(id, ReplicationSerDeHelper.parseTableCFsFromConfig(tableCfs)); } @@ -283,13 +283,15 @@ public class ReplicationAdmin implements Closeable { * @param id a short that identifies the cluster * @param tableCfs A map from tableName to column family names * @throws ReplicationException + * @throws IOException */ public void appendPeerTableCFs(String id, Map<TableName, ? extends Collection<String>> tableCfs) - throws ReplicationException { + throws ReplicationException, IOException { if (tableCfs == null) { throw new ReplicationException("tableCfs is null"); } - Map<TableName, List<String>> preTableCfs = this.replicationPeers.getPeerTableCFsConfig(id); + ReplicationPeerConfig peerConfig = admin.getReplicationPeerConfig(id); + Map<TableName, List<String>> preTableCfs = peerConfig.getTableCFsMap(); if (preTableCfs == null) { setPeerTableCFs(id, tableCfs); return; @@ -314,7 +316,7 @@ public class ReplicationAdmin implements Closeable { } } } - setPeerTableCFs(id, preTableCfs); + updatePeerConfig(id, peerConfig); } /** @@ -322,11 +324,13 @@ public class ReplicationAdmin implements Closeable { * @param id a short name that identifies the cluster * @param tableCf table-cfs config str * @throws ReplicationException + * @throws IOException * @deprecated as release of 2.0.0, and it will be removed in 3.0.0, * use {@link #removePeerTableCFs(String, Map)} instead. */ @Deprecated - public void removePeerTableCFs(String id, String tableCf) throws ReplicationException { + public void removePeerTableCFs(String id, String tableCf) throws ReplicationException, + IOException { removePeerTableCFs(id, ReplicationSerDeHelper.parseTableCFsFromConfig(tableCf)); } @@ -335,13 +339,15 @@ public class ReplicationAdmin implements Closeable { * @param id a short name that identifies the cluster * @param tableCfs A map from tableName to column family names * @throws ReplicationException + * @throws IOException */ public void removePeerTableCFs(String id, Map<TableName, ? extends Collection<String>> tableCfs) - throws ReplicationException { + throws ReplicationException, IOException { if (tableCfs == null) { throw new ReplicationException("tableCfs is null"); } - Map<TableName, List<String>> preTableCfs = this.replicationPeers.getPeerTableCFsConfig(id); + ReplicationPeerConfig peerConfig = admin.getReplicationPeerConfig(id); + Map<TableName, List<String>> preTableCfs = peerConfig.getTableCFsMap(); if (preTableCfs == null) { throw new ReplicationException("Table-Cfs for peer" + id + " is null"); } @@ -372,7 +378,7 @@ public class ReplicationAdmin implements Closeable { throw new ReplicationException("No table: " + table + " in table-cfs config of peer: " + id); } } - setPeerTableCFs(id, preTableCfs); + updatePeerConfig(id, peerConfig); } /** @@ -384,10 +390,10 @@ public class ReplicationAdmin implements Closeable { * families */ public void setPeerTableCFs(String id, Map<TableName, ? extends Collection<String>> tableCfs) - throws ReplicationException { - checkNamespacesAndTableCfsConfigConflict( - this.replicationPeers.getReplicationPeerConfig(id).getNamespaces(), tableCfs); - this.replicationPeers.setPeerTableCFsConfig(id, tableCfs); + throws IOException { + ReplicationPeerConfig peerConfig = getPeerConfig(id); + peerConfig.setTableCFsMap(tableCfs); + updatePeerConfig(id, peerConfig); } /** http://git-wip-us.apache.org/repos/asf/hbase/blob/0e486656/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java index 424d578..4231a82 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/RequestConverter.java @@ -115,7 +115,9 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos; import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest; import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest; import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest; +import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest; import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest; +import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; import org.apache.hadoop.hbase.util.Pair; @@ -1595,4 +1597,20 @@ public final class RequestConverter { builder.setPeerId(peerId); return builder.build(); } + + public static GetReplicationPeerConfigRequest buildGetReplicationPeerConfigRequest( + String peerId) { + GetReplicationPeerConfigRequest.Builder builder = GetReplicationPeerConfigRequest.newBuilder(); + builder.setPeerId(peerId); + return builder.build(); + } + + public static UpdateReplicationPeerConfigRequest buildUpdateReplicationPeerConfigRequest( + String peerId, ReplicationPeerConfig peerConfig) { + UpdateReplicationPeerConfigRequest.Builder builder = UpdateReplicationPeerConfigRequest + .newBuilder(); + builder.setPeerId(peerId); + builder.setPeerConfig(ReplicationSerDeHelper.convert(peerConfig)); + return builder.build(); + } } http://git-wip-us.apache.org/repos/asf/hbase/blob/0e486656/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java ---------------------------------------------------------------------- diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java index 1794a49..d56c534 100644 --- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java +++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/MasterProtos.java @@ -65991,7 +65991,7 @@ public final class MasterProtos { /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> */ - java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName> + java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName> getServerNameList(); /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> @@ -66004,7 +66004,7 @@ public final class MasterProtos { /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> */ - java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> + java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameOrBuilderList(); /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> @@ -66102,7 +66102,7 @@ public final class MasterProtos { /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> */ - public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> + public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameOrBuilderList() { return serverName_; } @@ -66409,7 +66409,7 @@ public final class MasterProtos { serverNameBuilder_ = null; serverName_ = other.serverName_; bitField0_ = (bitField0_ & ~0x00000001); - serverNameBuilder_ = + serverNameBuilder_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getServerNameFieldBuilder() : null; } else { @@ -66645,7 +66645,7 @@ public final class MasterProtos { /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> */ - public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> + public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameOrBuilderList() { if (serverNameBuilder_ != null) { return serverNameBuilder_.getMessageOrBuilderList(); @@ -66671,12 +66671,12 @@ public final class MasterProtos { /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> */ - public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder> + public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder> getServerNameBuilderList() { return getServerNameFieldBuilder().getBuilderList(); } private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3< - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> + org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameFieldBuilder() { if (serverNameBuilder_ == null) { serverNameBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3< @@ -66745,7 +66745,7 @@ public final class MasterProtos { /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> */ - java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName> + java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName> getServerNameList(); /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> @@ -66758,7 +66758,7 @@ public final class MasterProtos { /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> */ - java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> + java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameOrBuilderList(); /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> @@ -66856,7 +66856,7 @@ public final class MasterProtos { /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> */ - public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> + public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameOrBuilderList() { return serverName_; } @@ -67163,7 +67163,7 @@ public final class MasterProtos { serverNameBuilder_ = null; serverName_ = other.serverName_; bitField0_ = (bitField0_ & ~0x00000001); - serverNameBuilder_ = + serverNameBuilder_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getServerNameFieldBuilder() : null; } else { @@ -67399,7 +67399,7 @@ public final class MasterProtos { /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> */ - public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> + public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameOrBuilderList() { if (serverNameBuilder_ != null) { return serverNameBuilder_.getMessageOrBuilderList(); @@ -67425,12 +67425,12 @@ public final class MasterProtos { /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> */ - public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder> + public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder> getServerNameBuilderList() { return getServerNameFieldBuilder().getBuilderList(); } private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3< - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> + org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameFieldBuilder() { if (serverNameBuilder_ == null) { serverNameBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3< @@ -67872,7 +67872,7 @@ public final class MasterProtos { /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> */ - java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName> + java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName> getServerNameList(); /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> @@ -67885,7 +67885,7 @@ public final class MasterProtos { /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> */ - java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> + java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameOrBuilderList(); /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> @@ -67983,7 +67983,7 @@ public final class MasterProtos { /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> */ - public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> + public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameOrBuilderList() { return serverName_; } @@ -68290,7 +68290,7 @@ public final class MasterProtos { serverNameBuilder_ = null; serverName_ = other.serverName_; bitField0_ = (bitField0_ & ~0x00000001); - serverNameBuilder_ = + serverNameBuilder_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getServerNameFieldBuilder() : null; } else { @@ -68526,7 +68526,7 @@ public final class MasterProtos { /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> */ - public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> + public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameOrBuilderList() { if (serverNameBuilder_ != null) { return serverNameBuilder_.getMessageOrBuilderList(); @@ -68552,12 +68552,12 @@ public final class MasterProtos { /** * <code>repeated .hbase.pb.ServerName server_name = 1;</code> */ - public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder> + public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder> getServerNameBuilderList() { return getServerNameFieldBuilder().getBuilderList(); } private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3< - org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> + org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameFieldBuilder() { if (serverNameBuilder_ == null) { serverNameBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3< @@ -69775,7 +69775,31 @@ public final class MasterProtos { /** * <pre> - ** Returns a list of ServerNames marked as draining. + ** Return peer config for a replication peer + * </pre> + * + * <code>rpc GetReplicationPeerConfig(.hbase.pb.GetReplicationPeerConfigRequest) returns (.hbase.pb.GetReplicationPeerConfigResponse);</code> + */ + public abstract void getReplicationPeerConfig( + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest request, + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse> done); + + /** + * <pre> + ** Update peer config for a replication peer + * </pre> + * + * <code>rpc UpdateReplicationPeerConfig(.hbase.pb.UpdateReplicationPeerConfigRequest) returns (.hbase.pb.UpdateReplicationPeerConfigResponse);</code> + */ + public abstract void updateReplicationPeerConfig( + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest request, + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse> done); + + /** + * <pre> + ** Returns a list of ServerNames marked as draining. * </pre> * * <code>rpc listDrainingRegionServers(.hbase.pb.ListDrainingRegionServersRequest) returns (.hbase.pb.ListDrainingRegionServersResponse);</code> @@ -69787,7 +69811,7 @@ public final class MasterProtos { /** * <pre> - ** Mark a list of ServerNames as draining. + ** Mark a list of ServerNames as draining. * </pre> * * <code>rpc drainRegionServers(.hbase.pb.DrainRegionServersRequest) returns (.hbase.pb.DrainRegionServersResponse);</code> @@ -69799,7 +69823,7 @@ public final class MasterProtos { /** * <pre> - ** Unmark a list of ServerNames marked as draining. + ** Unmark a list of ServerNames marked as draining. * </pre> * * <code>rpc removeDrainFromRegionServers(.hbase.pb.RemoveDrainFromRegionServersRequest) returns (.hbase.pb.RemoveDrainFromRegionServersResponse);</code> @@ -70311,6 +70335,22 @@ public final class MasterProtos { } @java.lang.Override + public void getReplicationPeerConfig( + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest request, + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse> done) { + impl.getReplicationPeerConfig(controller, request, done); + } + + @java.lang.Override + public void updateReplicationPeerConfig( + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest request, + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse> done) { + impl.updateReplicationPeerConfig(controller, request, done); + } + + @java.lang.Override public void listDrainingRegionServers( org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersRequest request, @@ -70481,10 +70521,14 @@ public final class MasterProtos { case 61: return impl.disableReplicationPeer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest)request); case 62: - return impl.listDrainingRegionServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersRequest)request); + return impl.getReplicationPeerConfig(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest)request); case 63: - return impl.drainRegionServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest)request); + return impl.updateReplicationPeerConfig(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest)request); case 64: + return impl.listDrainingRegionServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersRequest)request); + case 65: + return impl.drainRegionServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest)request); + case 66: return impl.removeDrainFromRegionServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersRequest)request); default: throw new java.lang.AssertionError("Can't get here."); @@ -70625,10 +70669,14 @@ public final class MasterProtos { case 61: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest.getDefaultInstance(); case 62: - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersRequest.getDefaultInstance(); + return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest.getDefaultInstance(); case 63: - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest.getDefaultInstance(); + return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest.getDefaultInstance(); case 64: + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersRequest.getDefaultInstance(); + case 65: + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest.getDefaultInstance(); + case 66: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersRequest.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); @@ -70769,10 +70817,14 @@ public final class MasterProtos { case 61: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.getDefaultInstance(); case 62: - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersResponse.getDefaultInstance(); + return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.getDefaultInstance(); case 63: - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse.getDefaultInstance(); + return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.getDefaultInstance(); case 64: + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersResponse.getDefaultInstance(); + case 65: + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse.getDefaultInstance(); + case 66: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersResponse.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); @@ -71557,7 +71609,31 @@ public final class MasterProtos { /** * <pre> - ** Returns a list of ServerNames marked as draining. + ** Return peer config for a replication peer + * </pre> + * + * <code>rpc GetReplicationPeerConfig(.hbase.pb.GetReplicationPeerConfigRequest) returns (.hbase.pb.GetReplicationPeerConfigResponse);</code> + */ + public abstract void getReplicationPeerConfig( + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest request, + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse> done); + + /** + * <pre> + ** Update peer config for a replication peer + * </pre> + * + * <code>rpc UpdateReplicationPeerConfig(.hbase.pb.UpdateReplicationPeerConfigRequest) returns (.hbase.pb.UpdateReplicationPeerConfigResponse);</code> + */ + public abstract void updateReplicationPeerConfig( + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest request, + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse> done); + + /** + * <pre> + ** Returns a list of ServerNames marked as draining. * </pre> * * <code>rpc listDrainingRegionServers(.hbase.pb.ListDrainingRegionServersRequest) returns (.hbase.pb.ListDrainingRegionServersResponse);</code> @@ -71569,7 +71645,7 @@ public final class MasterProtos { /** * <pre> - ** Mark a list of ServerNames as draining. + ** Mark a list of ServerNames as draining. * </pre> * * <code>rpc drainRegionServers(.hbase.pb.DrainRegionServersRequest) returns (.hbase.pb.DrainRegionServersResponse);</code> @@ -71581,7 +71657,7 @@ public final class MasterProtos { /** * <pre> - ** Unmark a list of ServerNames marked as draining. + ** Unmark a list of ServerNames marked as draining. * </pre> * * <code>rpc removeDrainFromRegionServers(.hbase.pb.RemoveDrainFromRegionServersRequest) returns (.hbase.pb.RemoveDrainFromRegionServersResponse);</code> @@ -71924,16 +72000,26 @@ public final class MasterProtos { done)); return; case 62: + this.getReplicationPeerConfig(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest)request, + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse>specializeCallback( + done)); + return; + case 63: + this.updateReplicationPeerConfig(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest)request, + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse>specializeCallback( + done)); + return; + case 64: this.listDrainingRegionServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersResponse>specializeCallback( done)); return; - case 63: + case 65: this.drainRegionServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse>specializeCallback( done)); return; - case 64: + case 66: this.removeDrainFromRegionServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersRequest)request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersResponse>specializeCallback( done)); @@ -72077,10 +72163,14 @@ public final class MasterProtos { case 61: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest.getDefaultInstance(); case 62: - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersRequest.getDefaultInstance(); + return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest.getDefaultInstance(); case 63: - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest.getDefaultInstance(); + return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest.getDefaultInstance(); case 64: + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersRequest.getDefaultInstance(); + case 65: + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest.getDefaultInstance(); + case 66: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersRequest.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); @@ -72221,10 +72311,14 @@ public final class MasterProtos { case 61: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.getDefaultInstance(); case 62: - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersResponse.getDefaultInstance(); + return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.getDefaultInstance(); case 63: - return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse.getDefaultInstance(); + return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.getDefaultInstance(); case 64: + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersResponse.getDefaultInstance(); + case 65: + return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse.getDefaultInstance(); + case 66: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersResponse.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); @@ -73177,12 +73271,42 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.getDefaultInstance())); } + public void getReplicationPeerConfig( + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest request, + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse> done) { + channel.callMethod( + getDescriptor().getMethods().get(62), + controller, + request, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.getDefaultInstance(), + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( + done, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.class, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.getDefaultInstance())); + } + + public void updateReplicationPeerConfig( + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest request, + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse> done) { + channel.callMethod( + getDescriptor().getMethods().get(63), + controller, + request, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.getDefaultInstance(), + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback( + done, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.class, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.getDefaultInstance())); + } + public void listDrainingRegionServers( org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersResponse> done) { channel.callMethod( - getDescriptor().getMethods().get(62), + getDescriptor().getMethods().get(64), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersResponse.getDefaultInstance(), @@ -73197,7 +73321,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse> done) { channel.callMethod( - getDescriptor().getMethods().get(63), + getDescriptor().getMethods().get(65), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse.getDefaultInstance(), @@ -73212,7 +73336,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersRequest request, org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersResponse> done) { channel.callMethod( - getDescriptor().getMethods().get(64), + getDescriptor().getMethods().get(66), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersResponse.getDefaultInstance(), @@ -73539,6 +73663,16 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException; + public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse getReplicationPeerConfig( + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest request) + throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException; + + public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse updateReplicationPeerConfig( + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest request) + throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException; + public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersResponse listDrainingRegionServers( org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersRequest request) @@ -74306,12 +74440,36 @@ public final class MasterProtos { } + public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse getReplicationPeerConfig( + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest request) + throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { + return (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse) channel.callBlockingMethod( + getDescriptor().getMethods().get(62), + controller, + request, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.getDefaultInstance()); + } + + + public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse updateReplicationPeerConfig( + org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest request) + throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { + return (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse) channel.callBlockingMethod( + getDescriptor().getMethods().get(63), + controller, + request, + org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.getDefaultInstance()); + } + + public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersResponse listDrainingRegionServers( org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(62), + getDescriptor().getMethods().get(64), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDrainingRegionServersResponse.getDefaultInstance()); @@ -74323,7 +74481,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(63), + getDescriptor().getMethods().get(65), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DrainRegionServersResponse.getDefaultInstance()); @@ -74335,7 +74493,7 @@ public final class MasterProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersRequest request) throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersResponse) channel.callBlockingMethod( - getDescriptor().getMethods().get(64), + getDescriptor().getMethods().get(66), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RemoveDrainFromRegionServersResponse.getDefaultInstance()); @@ -74913,32 +75071,32 @@ public final class MasterProtos { internal_static_hbase_pb_SecurityCapabilitiesResponse_fieldAccessorTable; private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ListDrainingRegionServersRequest_descriptor; - private static final + private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ListDrainingRegionServersRequest_fieldAccessorTable; private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ListDrainingRegionServersResponse_descriptor; - private static final + private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ListDrainingRegionServersResponse_fieldAccessorTable; private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_DrainRegionServersRequest_descriptor; - private static final + private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_DrainRegionServersRequest_fieldAccessorTable; private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_DrainRegionServersResponse_descriptor; - private static final + private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_DrainRegionServersResponse_fieldAccessorTable; private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_RemoveDrainFromRegionServersRequest_descriptor; - private static final + private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_RemoveDrainFromRegionServersRequest_fieldAccessorTable; private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_RemoveDrainFromRegionServersResponse_descriptor; - private static final + private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_RemoveDrainFromRegionServersResponse_fieldAccessorTable; @@ -75161,7 +75319,7 @@ public final class MasterProtos { "\022)\n\013server_name\030\001 \003(\0132\024.hbase.pb.ServerN" + "ame\"&\n$RemoveDrainFromRegionServersRespo", "nse*(\n\020MasterSwitchType\022\t\n\005SPLIT\020\000\022\t\n\005ME" + - "RGE\020\0012\251/\n\rMasterService\022e\n\024GetSchemaAlte" + + "RGE\020\0012\2301\n\rMasterService\022e\n\024GetSchemaAlte" + "rStatus\022%.hbase.pb.GetSchemaAlterStatusR" + "equest\032&.hbase.pb.GetSchemaAlterStatusRe" + "sponse\022b\n\023GetTableDescriptors\022$.hbase.pb" + @@ -75304,17 +75462,23 @@ public final class MasterProtos { "onPeerResponse\022k\n\026DisableReplicationPeer" + "\022\'.hbase.pb.DisableReplicationPeerReques" + "t\032(.hbase.pb.DisableReplicationPeerRespo" + - "nse\022t\n\031listDrainingRegionServers\022*.hbase" + - ".pb.ListDrainingRegionServersRequest\032+.h" + - "base.pb.ListDrainingRegionServersRespons" + - "e\022_\n\022drainRegionServers\022#.hbase.pb.Drain" + - "RegionServersRequest\032$.hbase.pb.DrainReg" + - "ionServersResponse\022}\n\034removeDrainFromReg", - "ionServers\022-.hbase.pb.RemoveDrainFromReg" + - "ionServersRequest\032..hbase.pb.RemoveDrain" + - "FromRegionServersResponseBI\n1org.apache." + - "hadoop.hbase.shaded.protobuf.generatedB\014" + - "MasterProtosH\001\210\001\001\240\001\001" + "nse\022q\n\030GetReplicationPeerConfig\022).hbase." + + "pb.GetReplicationPeerConfigRequest\032*.hba" + + "se.pb.GetReplicationPeerConfigResponse\022z" + + "\n\033UpdateReplicationPeerConfig\022,.hbase.pb" + + ".UpdateReplicationPeerConfigRequest\032-.hb" + + "ase.pb.UpdateReplicationPeerConfigRespon", + "se\022t\n\031listDrainingRegionServers\022*.hbase." + + "pb.ListDrainingRegionServersRequest\032+.hb" + + "ase.pb.ListDrainingRegionServersResponse" + + "\022_\n\022drainRegionServers\022#.hbase.pb.DrainR" + + "egionServersRequest\032$.hbase.pb.DrainRegi" + + "onServersResponse\022}\n\034removeDrainFromRegi" + + "onServers\022-.hbase.pb.RemoveDrainFromRegi" + + "onServersRequest\032..hbase.pb.RemoveDrainF" + + "romRegionServersResponseBI\n1org.apache.h" + + "adoop.hbase.shaded.protobuf.generatedB\014M", + "asterProtosH\001\210\001\001\240\001\001" }; org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
