http://git-wip-us.apache.org/repos/asf/incubator-trafodion/blob/c74e3d62/core/sqf/src/seatrans/hbase-trx/src/main/java/org/apache/hadoop/hbase/coprocessor/transactional/generated/TrxRegionProtos.java ---------------------------------------------------------------------- diff --git a/core/sqf/src/seatrans/hbase-trx/src/main/java/org/apache/hadoop/hbase/coprocessor/transactional/generated/TrxRegionProtos.java b/core/sqf/src/seatrans/hbase-trx/src/main/java/org/apache/hadoop/hbase/coprocessor/transactional/generated/TrxRegionProtos.java index fa3f5a0..18c6c58 100755 --- a/core/sqf/src/seatrans/hbase-trx/src/main/java/org/apache/hadoop/hbase/coprocessor/transactional/generated/TrxRegionProtos.java +++ b/core/sqf/src/seatrans/hbase-trx/src/main/java/org/apache/hadoop/hbase/coprocessor/transactional/generated/TrxRegionProtos.java @@ -1,3 +1,24 @@ +// @@@ START COPYRIGHT @@ +// +// Licensed to the Apache Software Foundation (ASF) under one +// or more contributor license agreements. See the NOTICE file +// distributed with this work for additional information +// regarding copyright ownership. The ASF licenses this file +// to you under the Apache License, Version 2.0 (the +// "License"); you may not use this file except in compliance +// with the License. You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, +// software distributed under the License is distributed on an +// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +// KIND, either express or implied. See the License for the +// specific language governing permissions and limitations +// under the License. +// +// @@@ END COPYRIGHT @@@ + // Generated by the protocol buffer compiler. DO NOT EDIT! // source: TrxRegion.proto @@ -6052,15 +6073,15 @@ public final class TrxRegionProtos { */ long getTransactionId(); - // required int64 startEpoc = 3; + // required int64 startEpoch = 3; /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - boolean hasStartEpoc(); + boolean hasStartEpoch(); /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - long getStartEpoc(); + long getStartEpoch(); // required int32 participantNum = 4; /** @@ -6145,7 +6166,7 @@ public final class TrxRegionProtos { } case 24: { bitField0_ |= 0x00000004; - startEpoc_ = input.readInt64(); + startEpoch_ = input.readInt64(); break; } case 32: { @@ -6230,20 +6251,20 @@ public final class TrxRegionProtos { return transactionId_; } - // required int64 startEpoc = 3; - public static final int STARTEPOC_FIELD_NUMBER = 3; - private long startEpoc_; + // required int64 startEpoch = 3; + public static final int STARTEPOCH_FIELD_NUMBER = 3; + private long startEpoch_; /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - public boolean hasStartEpoc() { + public boolean hasStartEpoch() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - public long getStartEpoc() { - return startEpoc_; + public long getStartEpoch() { + return startEpoch_; } // required int32 participantNum = 4; @@ -6281,7 +6302,7 @@ public final class TrxRegionProtos { private void initFields() { regionName_ = com.google.protobuf.ByteString.EMPTY; transactionId_ = 0L; - startEpoc_ = 0L; + startEpoch_ = 0L; participantNum_ = 0; dropTableRecorded_ = false; } @@ -6298,7 +6319,7 @@ public final class TrxRegionProtos { memoizedIsInitialized = 0; return false; } - if (!hasStartEpoc()) { + if (!hasStartEpoch()) { memoizedIsInitialized = 0; return false; } @@ -6324,7 +6345,7 @@ public final class TrxRegionProtos { output.writeInt64(2, transactionId_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeInt64(3, startEpoc_); + output.writeInt64(3, startEpoch_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { output.writeInt32(4, participantNum_); @@ -6351,7 +6372,7 @@ public final class TrxRegionProtos { } if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream - .computeInt64Size(3, startEpoc_); + .computeInt64Size(3, startEpoch_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { size += com.google.protobuf.CodedOutputStream @@ -6481,7 +6502,7 @@ public final class TrxRegionProtos { bitField0_ = (bitField0_ & ~0x00000001); transactionId_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); - startEpoc_ = 0L; + startEpoch_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); participantNum_ = 0; bitField0_ = (bitField0_ & ~0x00000008); @@ -6526,7 +6547,7 @@ public final class TrxRegionProtos { if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000004; } - result.startEpoc_ = startEpoc_; + result.startEpoch_ = startEpoch_; if (((from_bitField0_ & 0x00000008) == 0x00000008)) { to_bitField0_ |= 0x00000008; } @@ -6557,8 +6578,8 @@ public final class TrxRegionProtos { if (other.hasTransactionId()) { setTransactionId(other.getTransactionId()); } - if (other.hasStartEpoc()) { - setStartEpoc(other.getStartEpoc()); + if (other.hasStartEpoch()) { + setStartEpoch(other.getStartEpoch()); } if (other.hasParticipantNum()) { setParticipantNum(other.getParticipantNum()); @@ -6579,7 +6600,7 @@ public final class TrxRegionProtos { return false; } - if (!hasStartEpoc()) { + if (!hasStartEpoch()) { return false; } @@ -6682,35 +6703,35 @@ public final class TrxRegionProtos { return this; } - // required int64 startEpoc = 3; - private long startEpoc_ ; + // required int64 startEpoch = 3; + private long startEpoch_ ; /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - public boolean hasStartEpoc() { + public boolean hasStartEpoch() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - public long getStartEpoc() { - return startEpoc_; + public long getStartEpoch() { + return startEpoch_; } /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - public Builder setStartEpoc(long value) { + public Builder setStartEpoch(long value) { bitField0_ |= 0x00000004; - startEpoc_ = value; + startEpoch_ = value; onChanged(); return this; } /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - public Builder clearStartEpoc() { + public Builder clearStartEpoch() { bitField0_ = (bitField0_ & ~0x00000004); - startEpoc_ = 0L; + startEpoch_ = 0L; onChanged(); return this; } @@ -7460,15 +7481,15 @@ public final class TrxRegionProtos { */ long getTransactionId(); - // required int64 startEpoc = 3; + // required int64 startEpoch = 3; /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - boolean hasStartEpoc(); + boolean hasStartEpoch(); /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - long getStartEpoc(); + long getStartEpoch(); // required int32 participantNum = 4; /** @@ -7546,7 +7567,7 @@ public final class TrxRegionProtos { } case 24: { bitField0_ |= 0x00000002; - startEpoc_ = input.readInt64(); + startEpoch_ = input.readInt64(); break; } case 32: { @@ -7636,20 +7657,20 @@ public final class TrxRegionProtos { return transactionId_; } - // required int64 startEpoc = 3; - public static final int STARTEPOC_FIELD_NUMBER = 3; - private long startEpoc_; + // required int64 startEpoch = 3; + public static final int STARTEPOCH_FIELD_NUMBER = 3; + private long startEpoch_; /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - public boolean hasStartEpoc() { + public boolean hasStartEpoch() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - public long getStartEpoc() { - return startEpoc_; + public long getStartEpoch() { + return startEpoch_; } // required int32 participantNum = 4; @@ -7671,7 +7692,7 @@ public final class TrxRegionProtos { private void initFields() { regionName_ = java.util.Collections.emptyList(); transactionId_ = 0L; - startEpoc_ = 0L; + startEpoch_ = 0L; participantNum_ = 0; } private byte memoizedIsInitialized = -1; @@ -7683,7 +7704,7 @@ public final class TrxRegionProtos { memoizedIsInitialized = 0; return false; } - if (!hasStartEpoc()) { + if (!hasStartEpoch()) { memoizedIsInitialized = 0; return false; } @@ -7705,7 +7726,7 @@ public final class TrxRegionProtos { output.writeInt64(2, transactionId_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeInt64(3, startEpoc_); + output.writeInt64(3, startEpoch_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { output.writeInt32(4, participantNum_); @@ -7734,7 +7755,7 @@ public final class TrxRegionProtos { } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream - .computeInt64Size(3, startEpoc_); + .computeInt64Size(3, startEpoch_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream @@ -7860,7 +7881,7 @@ public final class TrxRegionProtos { bitField0_ = (bitField0_ & ~0x00000001); transactionId_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); - startEpoc_ = 0L; + startEpoch_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); participantNum_ = 0; bitField0_ = (bitField0_ & ~0x00000008); @@ -7904,7 +7925,7 @@ public final class TrxRegionProtos { if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000002; } - result.startEpoc_ = startEpoc_; + result.startEpoch_ = startEpoch_; if (((from_bitField0_ & 0x00000008) == 0x00000008)) { to_bitField0_ |= 0x00000004; } @@ -7938,8 +7959,8 @@ public final class TrxRegionProtos { if (other.hasTransactionId()) { setTransactionId(other.getTransactionId()); } - if (other.hasStartEpoc()) { - setStartEpoc(other.getStartEpoc()); + if (other.hasStartEpoch()) { + setStartEpoch(other.getStartEpoch()); } if (other.hasParticipantNum()) { setParticipantNum(other.getParticipantNum()); @@ -7953,7 +7974,7 @@ public final class TrxRegionProtos { return false; } - if (!hasStartEpoc()) { + if (!hasStartEpoch()) { return false; } @@ -8088,35 +8109,35 @@ public final class TrxRegionProtos { return this; } - // required int64 startEpoc = 3; - private long startEpoc_ ; + // required int64 startEpoch = 3; + private long startEpoch_ ; /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - public boolean hasStartEpoc() { + public boolean hasStartEpoch() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - public long getStartEpoc() { - return startEpoc_; + public long getStartEpoch() { + return startEpoch_; } /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - public Builder setStartEpoc(long value) { + public Builder setStartEpoch(long value) { bitField0_ |= 0x00000004; - startEpoc_ = value; + startEpoch_ = value; onChanged(); return this; } /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - public Builder clearStartEpoc() { + public Builder clearStartEpoch() { bitField0_ = (bitField0_ & ~0x00000004); - startEpoc_ = 0L; + startEpoch_ = 0L; onChanged(); return this; } @@ -8926,15 +8947,15 @@ public final class TrxRegionProtos { */ long getTransactionId(); - // required int64 startEpoc = 3; + // required int64 startEpoch = 3; /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - boolean hasStartEpoc(); + boolean hasStartEpoch(); /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - long getStartEpoc(); + long getStartEpoch(); // required int64 commitId = 4; /** @@ -9019,7 +9040,7 @@ public final class TrxRegionProtos { } case 24: { bitField0_ |= 0x00000004; - startEpoc_ = input.readInt64(); + startEpoch_ = input.readInt64(); break; } case 32: { @@ -9104,20 +9125,20 @@ public final class TrxRegionProtos { return transactionId_; } - // required int64 startEpoc = 3; - public static final int STARTEPOC_FIELD_NUMBER = 3; - private long startEpoc_; + // required int64 startEpoch = 3; + public static final int STARTEPOCH_FIELD_NUMBER = 3; + private long startEpoch_; /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - public boolean hasStartEpoc() { + public boolean hasStartEpoch() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - public long getStartEpoc() { - return startEpoc_; + public long getStartEpoch() { + return startEpoch_; } // required int64 commitId = 4; @@ -9155,7 +9176,7 @@ public final class TrxRegionProtos { private void initFields() { regionName_ = com.google.protobuf.ByteString.EMPTY; transactionId_ = 0L; - startEpoc_ = 0L; + startEpoch_ = 0L; commitId_ = 0L; participantNum_ = 0; } @@ -9172,7 +9193,7 @@ public final class TrxRegionProtos { memoizedIsInitialized = 0; return false; } - if (!hasStartEpoc()) { + if (!hasStartEpoch()) { memoizedIsInitialized = 0; return false; } @@ -9198,7 +9219,7 @@ public final class TrxRegionProtos { output.writeInt64(2, transactionId_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeInt64(3, startEpoc_); + output.writeInt64(3, startEpoch_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { output.writeInt64(4, commitId_); @@ -9225,7 +9246,7 @@ public final class TrxRegionProtos { } if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream - .computeInt64Size(3, startEpoc_); + .computeInt64Size(3, startEpoch_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { size += com.google.protobuf.CodedOutputStream @@ -9355,7 +9376,7 @@ public final class TrxRegionProtos { bitField0_ = (bitField0_ & ~0x00000001); transactionId_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); - startEpoc_ = 0L; + startEpoch_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); commitId_ = 0L; bitField0_ = (bitField0_ & ~0x00000008); @@ -9400,7 +9421,7 @@ public final class TrxRegionProtos { if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000004; } - result.startEpoc_ = startEpoc_; + result.startEpoch_ = startEpoch_; if (((from_bitField0_ & 0x00000008) == 0x00000008)) { to_bitField0_ |= 0x00000008; } @@ -9431,8 +9452,8 @@ public final class TrxRegionProtos { if (other.hasTransactionId()) { setTransactionId(other.getTransactionId()); } - if (other.hasStartEpoc()) { - setStartEpoc(other.getStartEpoc()); + if (other.hasStartEpoch()) { + setStartEpoch(other.getStartEpoch()); } if (other.hasCommitId()) { setCommitId(other.getCommitId()); @@ -9453,7 +9474,7 @@ public final class TrxRegionProtos { return false; } - if (!hasStartEpoc()) { + if (!hasStartEpoch()) { return false; } @@ -9556,35 +9577,35 @@ public final class TrxRegionProtos { return this; } - // required int64 startEpoc = 3; - private long startEpoc_ ; + // required int64 startEpoch = 3; + private long startEpoch_ ; /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - public boolean hasStartEpoc() { + public boolean hasStartEpoch() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - public long getStartEpoc() { - return startEpoc_; + public long getStartEpoch() { + return startEpoch_; } /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - public Builder setStartEpoc(long value) { + public Builder setStartEpoch(long value) { bitField0_ |= 0x00000004; - startEpoc_ = value; + startEpoch_ = value; onChanged(); return this; } /** - * <code>required int64 startEpoc = 3;</code> + * <code>required int64 startEpoch = 3;</code> */ - public Builder clearStartEpoc() { + public Builder clearStartEpoch() { bitField0_ = (bitField0_ & ~0x00000004); - startEpoc_ = 0L; + startEpoch_ = 0L; onChanged(); return this; } @@ -25731,58 +25752,58 @@ public final class TrxRegionProtos { // @@protoc_insertion_point(class_scope:PutMultipleTransactionalResponse) } - public interface RecoveryRequestRequestOrBuilder + public interface PushEpochRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required int64 transactionId = 1; + // required bytes regionName = 1; /** - * <code>required int64 transactionId = 1;</code> + * <code>required bytes regionName = 1;</code> */ - boolean hasTransactionId(); + boolean hasRegionName(); /** - * <code>required int64 transactionId = 1;</code> + * <code>required bytes regionName = 1;</code> */ - long getTransactionId(); + com.google.protobuf.ByteString getRegionName(); - // required bytes regionName = 2; + // required int64 transactionId = 2; /** - * <code>required bytes regionName = 2;</code> + * <code>required int64 transactionId = 2;</code> */ - boolean hasRegionName(); + boolean hasTransactionId(); /** - * <code>required bytes regionName = 2;</code> + * <code>required int64 transactionId = 2;</code> */ - com.google.protobuf.ByteString getRegionName(); + long getTransactionId(); - // required int32 tmId = 3; + // required int64 epoch = 3; /** - * <code>required int32 tmId = 3;</code> + * <code>required int64 epoch = 3;</code> */ - boolean hasTmId(); + boolean hasEpoch(); /** - * <code>required int32 tmId = 3;</code> + * <code>required int64 epoch = 3;</code> */ - int getTmId(); + long getEpoch(); } /** - * Protobuf type {@code RecoveryRequestRequest} + * Protobuf type {@code PushEpochRequest} */ - public static final class RecoveryRequestRequest extends + public static final class PushEpochRequest extends com.google.protobuf.GeneratedMessage - implements RecoveryRequestRequestOrBuilder { - // Use RecoveryRequestRequest.newBuilder() to construct. - private RecoveryRequestRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + implements PushEpochRequestOrBuilder { + // Use PushEpochRequest.newBuilder() to construct. + private PushEpochRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private RecoveryRequestRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private PushEpochRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - private static final RecoveryRequestRequest defaultInstance; - public static RecoveryRequestRequest getDefaultInstance() { + private static final PushEpochRequest defaultInstance; + public static PushEpochRequest getDefaultInstance() { return defaultInstance; } - public RecoveryRequestRequest getDefaultInstanceForType() { + public PushEpochRequest getDefaultInstanceForType() { return defaultInstance; } @@ -25792,7 +25813,7 @@ public final class TrxRegionProtos { getUnknownFields() { return this.unknownFields; } - private RecoveryRequestRequest( + private PushEpochRequest( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { @@ -25815,19 +25836,19 @@ public final class TrxRegionProtos { } break; } - case 8: { + case 10: { bitField0_ |= 0x00000001; - transactionId_ = input.readInt64(); + regionName_ = input.readBytes(); break; } - case 18: { + case 16: { bitField0_ |= 0x00000002; - regionName_ = input.readBytes(); + transactionId_ = input.readInt64(); break; } case 24: { bitField0_ |= 0x00000004; - tmId_ = input.readInt32(); + epoch_ = input.readInt64(); break; } } @@ -25844,99 +25865,99 @@ public final class TrxRegionProtos { } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_RecoveryRequestRequest_descriptor; + return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_PushEpochRequest_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_RecoveryRequestRequest_fieldAccessorTable + return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_PushEpochRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest.class, org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest.Builder.class); + org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest.class, org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest.Builder.class); } - public static com.google.protobuf.Parser<RecoveryRequestRequest> PARSER = - new com.google.protobuf.AbstractParser<RecoveryRequestRequest>() { - public RecoveryRequestRequest parsePartialFrom( + public static com.google.protobuf.Parser<PushEpochRequest> PARSER = + new com.google.protobuf.AbstractParser<PushEpochRequest>() { + public PushEpochRequest parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - return new RecoveryRequestRequest(input, extensionRegistry); + return new PushEpochRequest(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser<RecoveryRequestRequest> getParserForType() { + public com.google.protobuf.Parser<PushEpochRequest> getParserForType() { return PARSER; } private int bitField0_; - // required int64 transactionId = 1; - public static final int TRANSACTIONID_FIELD_NUMBER = 1; - private long transactionId_; + // required bytes regionName = 1; + public static final int REGIONNAME_FIELD_NUMBER = 1; + private com.google.protobuf.ByteString regionName_; /** - * <code>required int64 transactionId = 1;</code> + * <code>required bytes regionName = 1;</code> */ - public boolean hasTransactionId() { + public boolean hasRegionName() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** - * <code>required int64 transactionId = 1;</code> + * <code>required bytes regionName = 1;</code> */ - public long getTransactionId() { - return transactionId_; + public com.google.protobuf.ByteString getRegionName() { + return regionName_; } - // required bytes regionName = 2; - public static final int REGIONNAME_FIELD_NUMBER = 2; - private com.google.protobuf.ByteString regionName_; + // required int64 transactionId = 2; + public static final int TRANSACTIONID_FIELD_NUMBER = 2; + private long transactionId_; /** - * <code>required bytes regionName = 2;</code> + * <code>required int64 transactionId = 2;</code> */ - public boolean hasRegionName() { + public boolean hasTransactionId() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** - * <code>required bytes regionName = 2;</code> + * <code>required int64 transactionId = 2;</code> */ - public com.google.protobuf.ByteString getRegionName() { - return regionName_; + public long getTransactionId() { + return transactionId_; } - // required int32 tmId = 3; - public static final int TMID_FIELD_NUMBER = 3; - private int tmId_; + // required int64 epoch = 3; + public static final int EPOCH_FIELD_NUMBER = 3; + private long epoch_; /** - * <code>required int32 tmId = 3;</code> + * <code>required int64 epoch = 3;</code> */ - public boolean hasTmId() { + public boolean hasEpoch() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** - * <code>required int32 tmId = 3;</code> + * <code>required int64 epoch = 3;</code> */ - public int getTmId() { - return tmId_; + public long getEpoch() { + return epoch_; } private void initFields() { - transactionId_ = 0L; regionName_ = com.google.protobuf.ByteString.EMPTY; - tmId_ = 0; + transactionId_ = 0L; + epoch_ = 0L; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; - if (!hasTransactionId()) { + if (!hasRegionName()) { memoizedIsInitialized = 0; return false; } - if (!hasRegionName()) { + if (!hasTransactionId()) { memoizedIsInitialized = 0; return false; } - if (!hasTmId()) { + if (!hasEpoch()) { memoizedIsInitialized = 0; return false; } @@ -25948,13 +25969,13 @@ public final class TrxRegionProtos { throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeInt64(1, transactionId_); + output.writeBytes(1, regionName_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, regionName_); + output.writeInt64(2, transactionId_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeInt32(3, tmId_); + output.writeInt64(3, epoch_); } getUnknownFields().writeTo(output); } @@ -25967,15 +25988,15 @@ public final class TrxRegionProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream - .computeInt64Size(1, transactionId_); + .computeBytesSize(1, regionName_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, regionName_); + .computeInt64Size(2, transactionId_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream - .computeInt32Size(3, tmId_); + .computeInt64Size(3, epoch_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; @@ -25989,53 +26010,53 @@ public final class TrxRegionProtos { return super.writeReplace(); } - public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseFrom( + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } - public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseFrom( + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } - public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseFrom(byte[] data) + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } - public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseFrom( + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } - public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } - public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseFrom( + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } - public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseDelimitedFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } - public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseDelimitedFrom( + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } - public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseFrom( + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } - public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseFrom( + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -26044,7 +26065,7 @@ public final class TrxRegionProtos { public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest prototype) { + public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @@ -26056,24 +26077,24 @@ public final class TrxRegionProtos { return builder; } /** - * Protobuf type {@code RecoveryRequestRequest} + * Protobuf type {@code PushEpochRequest} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> - implements org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequestOrBuilder { + implements org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequestOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_RecoveryRequestRequest_descriptor; + return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_PushEpochRequest_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_RecoveryRequestRequest_fieldAccessorTable + return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_PushEpochRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest.class, org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest.Builder.class); + org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest.class, org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest.Builder.class); } - // Construct using org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest.newBuilder() + // Construct using org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } @@ -26093,11 +26114,11 @@ public final class TrxRegionProtos { public Builder clear() { super.clear(); - transactionId_ = 0L; - bitField0_ = (bitField0_ & ~0x00000001); regionName_ = com.google.protobuf.ByteString.EMPTY; + bitField0_ = (bitField0_ & ~0x00000001); + transactionId_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); - tmId_ = 0; + epoch_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); return this; } @@ -26108,76 +26129,76 @@ public final class TrxRegionProtos { public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { - return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_RecoveryRequestRequest_descriptor; + return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_PushEpochRequest_descriptor; } - public org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest getDefaultInstanceForType() { - return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest.getDefaultInstance(); + public org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest getDefaultInstanceForType() { + return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest.getDefaultInstance(); } - public org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest build() { - org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest result = buildPartial(); + public org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest build() { + org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } - public org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest buildPartial() { - org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest result = new org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest(this); + public org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest buildPartial() { + org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest result = new org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } - result.transactionId_ = transactionId_; + result.regionName_ = regionName_; if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } - result.regionName_ = regionName_; + result.transactionId_ = transactionId_; if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000004; } - result.tmId_ = tmId_; + result.epoch_ = epoch_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest) { - return mergeFrom((org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest)other); + if (other instanceof org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest) { + return mergeFrom((org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest)other); } else { super.mergeFrom(other); return this; } } - public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest other) { - if (other == org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest.getDefaultInstance()) return this; - if (other.hasTransactionId()) { - setTransactionId(other.getTransactionId()); - } + public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest other) { + if (other == org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest.getDefaultInstance()) return this; if (other.hasRegionName()) { setRegionName(other.getRegionName()); } - if (other.hasTmId()) { - setTmId(other.getTmId()); + if (other.hasTransactionId()) { + setTransactionId(other.getTransactionId()); + } + if (other.hasEpoch()) { + setEpoch(other.getEpoch()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { - if (!hasTransactionId()) { + if (!hasRegionName()) { return false; } - if (!hasRegionName()) { + if (!hasTransactionId()) { return false; } - if (!hasTmId()) { + if (!hasEpoch()) { return false; } @@ -26188,11 +26209,11 @@ public final class TrxRegionProtos { com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parsedMessage = null; + org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest) e.getUnfinishedMessage(); + parsedMessage = (org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochRequest) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { @@ -26203,180 +26224,166 @@ public final class TrxRegionProtos { } private int bitField0_; - // required int64 transactionId = 1; - private long transactionId_ ; + // required bytes regionName = 1; + private com.google.protobuf.ByteString regionName_ = com.google.protobuf.ByteString.EMPTY; /** - * <code>required int64 transactionId = 1;</code> + * <code>required bytes regionName = 1;</code> */ - public boolean hasTransactionId() { + public boolean hasRegionName() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** - * <code>required int64 transactionId = 1;</code> + * <code>required bytes regionName = 1;</code> */ - public long getTransactionId() { - return transactionId_; + public com.google.protobuf.ByteString getRegionName() { + return regionName_; } /** - * <code>required int64 transactionId = 1;</code> + * <code>required bytes regionName = 1;</code> */ - public Builder setTransactionId(long value) { - bitField0_ |= 0x00000001; - transactionId_ = value; + public Builder setRegionName(com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000001; + regionName_ = value; onChanged(); return this; } /** - * <code>required int64 transactionId = 1;</code> + * <code>required bytes regionName = 1;</code> */ - public Builder clearTransactionId() { + public Builder clearRegionName() { bitField0_ = (bitField0_ & ~0x00000001); - transactionId_ = 0L; + regionName_ = getDefaultInstance().getRegionName(); onChanged(); return this; } - // required bytes regionName = 2; - private com.google.protobuf.ByteString regionName_ = com.google.protobuf.ByteString.EMPTY; + // required int64 transactionId = 2; + private long transactionId_ ; /** - * <code>required bytes regionName = 2;</code> + * <code>required int64 transactionId = 2;</code> */ - public boolean hasRegionName() { + public boolean hasTransactionId() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** - * <code>required bytes regionName = 2;</code> + * <code>required int64 transactionId = 2;</code> */ - public com.google.protobuf.ByteString getRegionName() { - return regionName_; + public long getTransactionId() { + return transactionId_; } /** - * <code>required bytes regionName = 2;</code> + * <code>required int64 transactionId = 2;</code> */ - public Builder setRegionName(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000002; - regionName_ = value; + public Builder setTransactionId(long value) { + bitField0_ |= 0x00000002; + transactionId_ = value; onChanged(); return this; } /** - * <code>required bytes regionName = 2;</code> + * <code>required int64 transactionId = 2;</code> */ - public Builder clearRegionName() { + public Builder clearTransactionId() { bitField0_ = (bitField0_ & ~0x00000002); - regionName_ = getDefaultInstance().getRegionName(); + transactionId_ = 0L; onChanged(); return this; } - // required int32 tmId = 3; - private int tmId_ ; + // required int64 epoch = 3; + private long epoch_ ; /** - * <code>required int32 tmId = 3;</code> + * <code>required int64 epoch = 3;</code> */ - public boolean hasTmId() { + public boolean hasEpoch() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** - * <code>required int32 tmId = 3;</code> + * <code>required int64 epoch = 3;</code> */ - public int getTmId() { - return tmId_; + public long getEpoch() { + return epoch_; } /** - * <code>required int32 tmId = 3;</code> + * <code>required int64 epoch = 3;</code> */ - public Builder setTmId(int value) { + public Builder setEpoch(long value) { bitField0_ |= 0x00000004; - tmId_ = value; + epoch_ = value; onChanged(); return this; } /** - * <code>required int32 tmId = 3;</code> + * <code>required int64 epoch = 3;</code> */ - public Builder clearTmId() { + public Builder clearEpoch() { bitField0_ = (bitField0_ & ~0x00000004); - tmId_ = 0; + epoch_ = 0L; onChanged(); return this; } - // @@protoc_insertion_point(builder_scope:RecoveryRequestRequest) + // @@protoc_insertion_point(builder_scope:PushEpochRequest) } static { - defaultInstance = new RecoveryRequestRequest(true); + defaultInstance = new PushEpochRequest(true); defaultInstance.initFields(); } - // @@protoc_insertion_point(class_scope:RecoveryRequestRequest) + // @@protoc_insertion_point(class_scope:PushEpochRequest) } - public interface RecoveryRequestResponseOrBuilder + public interface PushEpochResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // repeated int64 result = 1; - /** - * <code>repeated int64 result = 1;</code> - */ - java.util.List<java.lang.Long> getResultList(); - /** - * <code>repeated int64 result = 1;</code> - */ - int getResultCount(); - /** - * <code>repeated int64 result = 1;</code> - */ - long getResult(int index); - - // optional string exception = 2; + // optional string exception = 1; /** - * <code>optional string exception = 2;</code> + * <code>optional string exception = 1;</code> */ boolean hasException(); /** - * <code>optional string exception = 2;</code> + * <code>optional string exception = 1;</code> */ java.lang.String getException(); /** - * <code>optional string exception = 2;</code> + * <code>optional string exception = 1;</code> */ com.google.protobuf.ByteString getExceptionBytes(); - // optional bool hasException = 3; + // optional bool hasException = 2; /** - * <code>optional bool hasException = 3;</code> + * <code>optional bool hasException = 2;</code> */ boolean hasHasException(); /** - * <code>optional bool hasException = 3;</code> + * <code>optional bool hasException = 2;</code> */ boolean getHasException(); } /** - * Protobuf type {@code RecoveryRequestResponse} + * Protobuf type {@code PushEpochResponse} */ - public static final class RecoveryRequestResponse extends + public static final class PushEpochResponse extends com.google.protobuf.GeneratedMessage - implements RecoveryRequestResponseOrBuilder { - // Use RecoveryRequestResponse.newBuilder() to construct. - private RecoveryRequestResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + implements PushEpochResponseOrBuilder { + // Use PushEpochResponse.newBuilder() to construct. + private PushEpochResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private RecoveryRequestResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private PushEpochResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - private static final RecoveryRequestResponse defaultInstance; - public static RecoveryRequestResponse getDefaultInstance() { + private static final PushEpochResponse defaultInstance; + public static PushEpochResponse getDefaultInstance() { return defaultInstance; } - public RecoveryRequestResponse getDefaultInstanceForType() { + public PushEpochResponse getDefaultInstanceForType() { return defaultInstance; } @@ -26386,7 +26393,1158 @@ public final class TrxRegionProtos { getUnknownFields() { return this.unknownFields; } - private RecoveryRequestResponse( + private PushEpochResponse( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + initFields(); + int mutable_bitField0_ = 0; + com.google.protobuf.UnknownFieldSet.Builder unknownFields = + com.google.protobuf.UnknownFieldSet.newBuilder(); + try { + boolean done = false; + while (!done) { + int tag = input.readTag(); + switch (tag) { + case 0: + done = true; + break; + default: { + if (!parseUnknownField(input, unknownFields, + extensionRegistry, tag)) { + done = true; + } + break; + } + case 10: { + bitField0_ |= 0x00000001; + exception_ = input.readBytes(); + break; + } + case 16: { + bitField0_ |= 0x00000002; + hasException_ = input.readBool(); + break; + } + } + } + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + throw e.setUnfinishedMessage(this); + } catch (java.io.IOException e) { + throw new com.google.protobuf.InvalidProtocolBufferException( + e.getMessage()).setUnfinishedMessage(this); + } finally { + this.unknownFields = unknownFields.build(); + makeExtensionsImmutable(); + } + } + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_PushEpochResponse_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_PushEpochResponse_fieldAccessorTable + .ensureFieldAccessorsInitialized( + org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse.class, org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse.Builder.class); + } + + public static com.google.protobuf.Parser<PushEpochResponse> PARSER = + new com.google.protobuf.AbstractParser<PushEpochResponse>() { + public PushEpochResponse parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new PushEpochResponse(input, extensionRegistry); + } + }; + + @java.lang.Override + public com.google.protobuf.Parser<PushEpochResponse> getParserForType() { + return PARSER; + } + + private int bitField0_; + // optional string exception = 1; + public static final int EXCEPTION_FIELD_NUMBER = 1; + private java.lang.Object exception_; + /** + * <code>optional string exception = 1;</code> + */ + public boolean hasException() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + /** + * <code>optional string exception = 1;</code> + */ + public java.lang.String getException() { + java.lang.Object ref = exception_; + if (ref instanceof java.lang.String) { + return (java.lang.String) ref; + } else { + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + exception_ = s; + } + return s; + } + } + /** + * <code>optional string exception = 1;</code> + */ + public com.google.protobuf.ByteString + getExceptionBytes() { + java.lang.Object ref = exception_; + if (ref instanceof java.lang.String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8( + (java.lang.String) ref); + exception_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + + // optional bool hasException = 2; + public static final int HASEXCEPTION_FIELD_NUMBER = 2; + private boolean hasException_; + /** + * <code>optional bool hasException = 2;</code> + */ + public boolean hasHasException() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + /** + * <code>optional bool hasException = 2;</code> + */ + public boolean getHasException() { + return hasException_; + } + + private void initFields() { + exception_ = ""; + hasException_ = false; + } + private byte memoizedIsInitialized = -1; + public final boolean isInitialized() { + byte isInitialized = memoizedIsInitialized; + if (isInitialized != -1) return isInitialized == 1; + + memoizedIsInitialized = 1; + return true; + } + + public void writeTo(com.google.protobuf.CodedOutputStream output) + throws java.io.IOException { + getSerializedSize(); + if (((bitField0_ & 0x00000001) == 0x00000001)) { + output.writeBytes(1, getExceptionBytes()); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + output.writeBool(2, hasException_); + } + getUnknownFields().writeTo(output); + } + + private int memoizedSerializedSize = -1; + public int getSerializedSize() { + int size = memoizedSerializedSize; + if (size != -1) return size; + + size = 0; + if (((bitField0_ & 0x00000001) == 0x00000001)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(1, getExceptionBytes()); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + size += com.google.protobuf.CodedOutputStream + .computeBoolSize(2, hasException_); + } + size += getUnknownFields().getSerializedSize(); + memoizedSerializedSize = size; + return size; + } + + private static final long serialVersionUID = 0L; + @java.lang.Override + protected java.lang.Object writeReplace() + throws java.io.ObjectStreamException { + return super.writeReplace(); + } + + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse parseFrom(byte[] data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse parseFrom( + byte[] data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse parseFrom(java.io.InputStream input) + throws java.io.IOException { + return PARSER.parseFrom(input); + } + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse parseFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseFrom(input, extensionRegistry); + } + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse parseDelimitedFrom(java.io.InputStream input) + throws java.io.IOException { + return PARSER.parseDelimitedFrom(input); + } + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse parseDelimitedFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseDelimitedFrom(input, extensionRegistry); + } + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse parseFrom( + com.google.protobuf.CodedInputStream input) + throws java.io.IOException { + return PARSER.parseFrom(input); + } + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse parseFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseFrom(input, extensionRegistry); + } + + public static Builder newBuilder() { return Builder.create(); } + public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse prototype) { + return newBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { return newBuilder(this); } + + @java.lang.Override + protected Builder newBuilderForType( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + Builder builder = new Builder(parent); + return builder; + } + /** + * Protobuf type {@code PushEpochResponse} + */ + public static final class Builder extends + com.google.protobuf.GeneratedMessage.Builder<Builder> + implements org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponseOrBuilder { + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_PushEpochResponse_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_PushEpochResponse_fieldAccessorTable + .ensureFieldAccessorsInitialized( + org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse.class, org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse.Builder.class); + } + + // Construct using org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse.newBuilder() + private Builder() { + maybeForceBuilderInitialization(); + } + + private Builder( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + super(parent); + maybeForceBuilderInitialization(); + } + private void maybeForceBuilderInitialization() { + if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + } + } + private static Builder create() { + return new Builder(); + } + + public Builder clear() { + super.clear(); + exception_ = ""; + bitField0_ = (bitField0_ & ~0x00000001); + hasException_ = false; + bitField0_ = (bitField0_ & ~0x00000002); + return this; + } + + public Builder clone() { + return create().mergeFrom(buildPartial()); + } + + public com.google.protobuf.Descriptors.Descriptor + getDescriptorForType() { + return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_PushEpochResponse_descriptor; + } + + public org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse getDefaultInstanceForType() { + return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse.getDefaultInstance(); + } + + public org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse build() { + org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException(result); + } + return result; + } + + public org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse buildPartial() { + org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse result = new org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse(this); + int from_bitField0_ = bitField0_; + int to_bitField0_ = 0; + if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + to_bitField0_ |= 0x00000001; + } + result.exception_ = exception_; + if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + to_bitField0_ |= 0x00000002; + } + result.hasException_ = hasException_; + result.bitField0_ = to_bitField0_; + onBuilt(); + return result; + } + + public Builder mergeFrom(com.google.protobuf.Message other) { + if (other instanceof org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse) { + return mergeFrom((org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse)other); + } else { + super.mergeFrom(other); + return this; + } + } + + public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse other) { + if (other == org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse.getDefaultInstance()) return this; + if (other.hasException()) { + bitField0_ |= 0x00000001; + exception_ = other.exception_; + onChanged(); + } + if (other.hasHasException()) { + setHasException(other.getHasException()); + } + this.mergeUnknownFields(other.getUnknownFields()); + return this; + } + + public final boolean isInitialized() { + return true; + } + + public Builder mergeFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse parsedMessage = null; + try { + parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + parsedMessage = (org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.PushEpochResponse) e.getUnfinishedMessage(); + throw e; + } finally { + if (parsedMessage != null) { + mergeFrom(parsedMessage); + } + } + return this; + } + private int bitField0_; + + // optional string exception = 1; + private java.lang.Object exception_ = ""; + /** + * <code>optional string exception = 1;</code> + */ + public boolean hasException() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + /** + * <code>optional string exception = 1;</code> + */ + public java.lang.String getException() { + java.lang.Object ref = exception_; + if (!(ref instanceof java.lang.String)) { + java.lang.String s = ((com.google.protobuf.ByteString) ref) + .toStringUtf8(); + exception_ = s; + return s; + } else { + return (java.lang.String) ref; + } + } + /** + * <code>optional string exception = 1;</code> + */ + public com.google.protobuf.ByteString + getExceptionBytes() { + java.lang.Object ref = exception_; + if (ref instanceof String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8( + (java.lang.String) ref); + exception_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + /** + * <code>optional string exception = 1;</code> + */ + public Builder setException( + java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000001; + exception_ = value; + onChanged(); + return this; + } + /** + * <code>optional string exception = 1;</code> + */ + public Builder clearException() { + bitField0_ = (bitField0_ & ~0x00000001); + exception_ = getDefaultInstance().getException(); + onChanged(); + return this; + } + /** + * <code>optional string exception = 1;</code> + */ + public Builder setExceptionBytes( + com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000001; + exception_ = value; + onChanged(); + return this; + } + + // optional bool hasException = 2; + private boolean hasException_ ; + /** + * <code>optional bool hasException = 2;</code> + */ + public boolean hasHasException() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + /** + * <code>optional bool hasException = 2;</code> + */ + public boolean getHasException() { + return hasException_; + } + /** + * <code>optional bool hasException = 2;</code> + */ + public Builder setHasException(boolean value) { + bitField0_ |= 0x00000002; + hasException_ = value; + onChanged(); + return this; + } + /** + * <code>optional bool hasException = 2;</code> + */ + public Builder clearHasException() { + bitField0_ = (bitField0_ & ~0x00000002); + hasException_ = false; + onChanged(); + return this; + } + + // @@protoc_insertion_point(builder_scope:PushEpochResponse) + } + + static { + defaultInstance = new PushEpochResponse(true); + defaultInstance.initFields(); + } + + // @@protoc_insertion_point(class_scope:PushEpochResponse) + } + + public interface RecoveryRequestRequestOrBuilder + extends com.google.protobuf.MessageOrBuilder { + + // required int64 transactionId = 1; + /** + * <code>required int64 transactionId = 1;</code> + */ + boolean hasTransactionId(); + /** + * <code>required int64 transactionId = 1;</code> + */ + long getTransactionId(); + + // required bytes regionName = 2; + /** + * <code>required bytes regionName = 2;</code> + */ + boolean hasRegionName(); + /** + * <code>required bytes regionName = 2;</code> + */ + com.google.protobuf.ByteString getRegionName(); + + // required int32 tmId = 3; + /** + * <code>required int32 tmId = 3;</code> + */ + boolean hasTmId(); + /** + * <code>required int32 tmId = 3;</code> + */ + int getTmId(); + } + /** + * Protobuf type {@code RecoveryRequestRequest} + */ + public static final class RecoveryRequestRequest extends + com.google.protobuf.GeneratedMessage + implements RecoveryRequestRequestOrBuilder { + // Use RecoveryRequestRequest.newBuilder() to construct. + private RecoveryRequestRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + super(builder); + this.unknownFields = builder.getUnknownFields(); + } + private RecoveryRequestRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + + private static final RecoveryRequestRequest defaultInstance; + public static RecoveryRequestRequest getDefaultInstance() { + return defaultInstance; + } + + public RecoveryRequestRequest getDefaultInstanceForType() { + return defaultInstance; + } + + private final com.google.protobuf.UnknownFieldSet unknownFields; + @java.lang.Override + public final com.google.protobuf.UnknownFieldSet + getUnknownFields() { + return this.unknownFields; + } + private RecoveryRequestRequest( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + initFields(); + int mutable_bitField0_ = 0; + com.google.protobuf.UnknownFieldSet.Builder unknownFields = + com.google.protobuf.UnknownFieldSet.newBuilder(); + try { + boolean done = false; + while (!done) { + int tag = input.readTag(); + switch (tag) { + case 0: + done = true; + break; + default: { + if (!parseUnknownField(input, unknownFields, + extensionRegistry, tag)) { + done = true; + } + break; + } + case 8: { + bitField0_ |= 0x00000001; + transactionId_ = input.readInt64(); + break; + } + case 18: { + bitField0_ |= 0x00000002; + regionName_ = input.readBytes(); + break; + } + case 24: { + bitField0_ |= 0x00000004; + tmId_ = input.readInt32(); + break; + } + } + } + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + throw e.setUnfinishedMessage(this); + } catch (java.io.IOException e) { + throw new com.google.protobuf.InvalidProtocolBufferException( + e.getMessage()).setUnfinishedMessage(this); + } finally { + this.unknownFields = unknownFields.build(); + makeExtensionsImmutable(); + } + } + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_RecoveryRequestRequest_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_RecoveryRequestRequest_fieldAccessorTable + .ensureFieldAccessorsInitialized( + org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest.class, org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest.Builder.class); + } + + public static com.google.protobuf.Parser<RecoveryRequestRequest> PARSER = + new com.google.protobuf.AbstractParser<RecoveryRequestRequest>() { + public RecoveryRequestRequest parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new RecoveryRequestRequest(input, extensionRegistry); + } + }; + + @java.lang.Override + public com.google.protobuf.Parser<RecoveryRequestRequest> getParserForType() { + return PARSER; + } + + private int bitField0_; + // required int64 transactionId = 1; + public static final int TRANSACTIONID_FIELD_NUMBER = 1; + private long transactionId_; + /** + * <code>required int64 transactionId = 1;</code> + */ + public boolean hasTransactionId() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + /** + * <code>required int64 transactionId = 1;</code> + */ + public long getTransactionId() { + return transactionId_; + } + + // required bytes regionName = 2; + public static final int REGIONNAME_FIELD_NUMBER = 2; + private com.google.protobuf.ByteString regionName_; + /** + * <code>required bytes regionName = 2;</code> + */ + public boolean hasRegionName() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + /** + * <code>required bytes regionName = 2;</code> + */ + public com.google.protobuf.ByteString getRegionName() { + return regionName_; + } + + // required int32 tmId = 3; + public static final int TMID_FIELD_NUMBER = 3; + private int tmId_; + /** + * <code>required int32 tmId = 3;</code> + */ + public boolean hasTmId() { + return ((bitField0_ & 0x00000004) == 0x00000004); + } + /** + * <code>required int32 tmId = 3;</code> + */ + public int getTmId() { + return tmId_; + } + + private void initFields() { + transactionId_ = 0L; + regionName_ = com.google.protobuf.ByteString.EMPTY; + tmId_ = 0; + } + private byte memoizedIsInitialized = -1; + public final boolean isInitialized() { + byte isInitialized = memoizedIsInitialized; + if (isInitialized != -1) return isInitialized == 1; + + if (!hasTransactionId()) { + memoizedIsInitialized = 0; + return false; + } + if (!hasRegionName()) { + memoizedIsInitialized = 0; + return false; + } + if (!hasTmId()) { + memoizedIsInitialized = 0; + return false; + } + memoizedIsInitialized = 1; + return true; + } + + public void writeTo(com.google.protobuf.CodedOutputStream output) + throws java.io.IOException { + getSerializedSize(); + if (((bitField0_ & 0x00000001) == 0x00000001)) { + output.writeInt64(1, transactionId_); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + output.writeBytes(2, regionName_); + } + if (((bitField0_ & 0x00000004) == 0x00000004)) { + output.writeInt32(3, tmId_); + } + getUnknownFields().writeTo(output); + } + + private int memoizedSerializedSize = -1; + public int getSerializedSize() { + int size = memoizedSerializedSize; + if (size != -1) return size; + + size = 0; + if (((bitField0_ & 0x00000001) == 0x00000001)) { + size += com.google.protobuf.CodedOutputStream + .computeInt64Size(1, transactionId_); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(2, regionName_); + } + if (((bitField0_ & 0x00000004) == 0x00000004)) { + size += com.google.protobuf.CodedOutputStream + .computeInt32Size(3, tmId_); + } + size += getUnknownFields().getSerializedSize(); + memoizedSerializedSize = size; + return size; + } + + private static final long serialVersionUID = 0L; + @java.lang.Override + protected java.lang.Object writeReplace() + throws java.io.ObjectStreamException { + return super.writeReplace(); + } + + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseFrom(byte[] data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseFrom( + byte[] data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseFrom(java.io.InputStream input) + throws java.io.IOException { + return PARSER.parseFrom(input); + } + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseFrom(input, extensionRegistry); + } + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseDelimitedFrom(java.io.InputStream input) + throws java.io.IOException { + return PARSER.parseDelimitedFrom(input); + } + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseDelimitedFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseDelimitedFrom(input, extensionRegistry); + } + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseFrom( + com.google.protobuf.CodedInputStream input) + throws java.io.IOException { + return PARSER.parseFrom(input); + } + public static org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest parseFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseFrom(input, extensionRegistry); + } + + public static Builder newBuilder() { return Builder.create(); } + public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest prototype) { + return newBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { return newBuilder(this); } + + @java.lang.Override + protected Builder newBuilderForType( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + Builder builder = new Builder(parent); + return builder; + } + /** + * Protobuf type {@code RecoveryRequestRequest} + */ + public static final class Builder extends + com.google.protobuf.GeneratedMessage.Builder<Builder> + implements org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequestOrBuilder { + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_RecoveryRequestRequest_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_RecoveryRequestRequest_fieldAccessorTable + .ensureFieldAccessorsInitialized( + org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest.class, org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest.Builder.class); + } + + // Construct using org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest.newBuilder() + private Builder() { + maybeForceBuilderInitialization(); + } + + private Builder( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + super(parent); + maybeForceBuilderInitialization(); + } + private void maybeForceBuilderInitialization() { + if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + } + } + private static Builder create() { + return new Builder(); + } + + public Builder clear() { + super.clear(); + transactionId_ = 0L; + bitField0_ = (bitField0_ & ~0x00000001); + regionName_ = com.google.protobuf.ByteString.EMPTY; + bitField0_ = (bitField0_ & ~0x00000002); + tmId_ = 0; + bitField0_ = (bitField0_ & ~0x00000004); + return this; + } + + public Builder clone() { + return create().mergeFrom(buildPartial()); + } + + public com.google.protobuf.Descriptors.Descriptor + getDescriptorForType() { + return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.internal_static_RecoveryRequestRequest_descriptor; + } + + public org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest getDefaultInstanceForType() { + return org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest.getDefaultInstance(); + } + + public org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest build() { + org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException(result); + } + return result; + } + + public org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest buildPartial() { + org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest result = new org.apache.hadoop.hbase.coprocessor.transactional.generated.TrxRegionProtos.RecoveryRequestRequest(this); + int from_bitField0_ = bitField0_; + int to_bitField0_ = 0; + if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + to_bitField0_ |= 0x00000001; + } + result.transactionId_ = transactionId_; + if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + to_bitField0_ |= 0x00000002; + } + result.regionName_ = regionName_; + if (((from_bi
<TRUNCATED>
