http://git-wip-us.apache.org/repos/asf/hbase/blob/3cc5d190/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java
----------------------------------------------------------------------
diff --git
a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java
b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java
index dd55599..2f5322d 100644
---
a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java
+++
b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/HBaseProtos.java
@@ -2404,6 +2404,1576 @@ public final class HBaseProtos {
// @@protoc_insertion_point(class_scope:TableSchema)
}
+ public interface TableStateOrBuilder
+ extends com.google.protobuf.MessageOrBuilder {
+
+ // required .TableState.State state = 1;
+ /**
+ * <code>required .TableState.State state = 1;</code>
+ *
+ * <pre>
+ * This is the table's state.
+ * </pre>
+ */
+ boolean hasState();
+ /**
+ * <code>required .TableState.State state = 1;</code>
+ *
+ * <pre>
+ * This is the table's state.
+ * </pre>
+ */
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State
getState();
+
+ // required .TableName table = 2;
+ /**
+ * <code>required .TableName table = 2;</code>
+ */
+ boolean hasTable();
+ /**
+ * <code>required .TableName table = 2;</code>
+ */
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName
getTable();
+ /**
+ * <code>required .TableName table = 2;</code>
+ */
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder
getTableOrBuilder();
+
+ // optional uint64 timestamp = 3;
+ /**
+ * <code>optional uint64 timestamp = 3;</code>
+ */
+ boolean hasTimestamp();
+ /**
+ * <code>optional uint64 timestamp = 3;</code>
+ */
+ long getTimestamp();
+ }
+ /**
+ * Protobuf type {@code TableState}
+ *
+ * <pre>
+ ** Denotes state of the table
+ * </pre>
+ */
+ public static final class TableState extends
+ com.google.protobuf.GeneratedMessage
+ implements TableStateOrBuilder {
+ // Use TableState.newBuilder() to construct.
+ private TableState(com.google.protobuf.GeneratedMessage.Builder<?>
builder) {
+ super(builder);
+ this.unknownFields = builder.getUnknownFields();
+ }
+ private TableState(boolean noInit) { this.unknownFields =
com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
+ private static final TableState defaultInstance;
+ public static TableState getDefaultInstance() {
+ return defaultInstance;
+ }
+
+ public TableState getDefaultInstanceForType() {
+ return defaultInstance;
+ }
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private TableState(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 8: {
+ int rawValue = input.readEnum();
+
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State value =
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State.valueOf(rawValue);
+ if (value == null) {
+ unknownFields.mergeVarintField(1, rawValue);
+ } else {
+ bitField0_ |= 0x00000001;
+ state_ = value;
+ }
+ break;
+ }
+ case 18: {
+
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder
subBuilder = null;
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ subBuilder = table_.toBuilder();
+ }
+ table_ =
input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER,
extensionRegistry);
+ if (subBuilder != null) {
+ subBuilder.mergeFrom(table_);
+ table_ = subBuilder.buildPartial();
+ }
+ bitField0_ |= 0x00000002;
+ break;
+ }
+ case 24: {
+ bitField0_ |= 0x00000004;
+ timestamp_ = input.readUInt64();
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TableState_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TableState_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.class,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<TableState> PARSER =
+ new com.google.protobuf.AbstractParser<TableState>() {
+ public TableState parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new TableState(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<TableState> getParserForType() {
+ return PARSER;
+ }
+
+ /**
+ * Protobuf enum {@code TableState.State}
+ *
+ * <pre>
+ * Table's current state
+ * </pre>
+ */
+ public enum State
+ implements com.google.protobuf.ProtocolMessageEnum {
+ /**
+ * <code>ENABLED = 0;</code>
+ */
+ ENABLED(0, 0),
+ /**
+ * <code>DISABLED = 1;</code>
+ */
+ DISABLED(1, 1),
+ /**
+ * <code>DISABLING = 2;</code>
+ */
+ DISABLING(2, 2),
+ /**
+ * <code>ENABLING = 3;</code>
+ */
+ ENABLING(3, 3),
+ ;
+
+ /**
+ * <code>ENABLED = 0;</code>
+ */
+ public static final int ENABLED_VALUE = 0;
+ /**
+ * <code>DISABLED = 1;</code>
+ */
+ public static final int DISABLED_VALUE = 1;
+ /**
+ * <code>DISABLING = 2;</code>
+ */
+ public static final int DISABLING_VALUE = 2;
+ /**
+ * <code>ENABLING = 3;</code>
+ */
+ public static final int ENABLING_VALUE = 3;
+
+
+ public final int getNumber() { return value; }
+
+ public static State valueOf(int value) {
+ switch (value) {
+ case 0: return ENABLED;
+ case 1: return DISABLED;
+ case 2: return DISABLING;
+ case 3: return ENABLING;
+ default: return null;
+ }
+ }
+
+ public static com.google.protobuf.Internal.EnumLiteMap<State>
+ internalGetValueMap() {
+ return internalValueMap;
+ }
+ private static com.google.protobuf.Internal.EnumLiteMap<State>
+ internalValueMap =
+ new com.google.protobuf.Internal.EnumLiteMap<State>() {
+ public State findValueByNumber(int number) {
+ return State.valueOf(number);
+ }
+ };
+
+ public final com.google.protobuf.Descriptors.EnumValueDescriptor
+ getValueDescriptor() {
+ return getDescriptor().getValues().get(index);
+ }
+ public final com.google.protobuf.Descriptors.EnumDescriptor
+ getDescriptorForType() {
+ return getDescriptor();
+ }
+ public static final com.google.protobuf.Descriptors.EnumDescriptor
+ getDescriptor() {
+ return
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.getDescriptor().getEnumTypes().get(0);
+ }
+
+ private static final State[] VALUES = values();
+
+ public static State valueOf(
+ com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
+ if (desc.getType() != getDescriptor()) {
+ throw new java.lang.IllegalArgumentException(
+ "EnumValueDescriptor is not for this type.");
+ }
+ return VALUES[desc.getIndex()];
+ }
+
+ private final int index;
+ private final int value;
+
+ private State(int index, int value) {
+ this.index = index;
+ this.value = value;
+ }
+
+ // @@protoc_insertion_point(enum_scope:TableState.State)
+ }
+
+ private int bitField0_;
+ // required .TableState.State state = 1;
+ public static final int STATE_FIELD_NUMBER = 1;
+ private
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State state_;
+ /**
+ * <code>required .TableState.State state = 1;</code>
+ *
+ * <pre>
+ * This is the table's state.
+ * </pre>
+ */
+ public boolean hasState() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * <code>required .TableState.State state = 1;</code>
+ *
+ * <pre>
+ * This is the table's state.
+ * </pre>
+ */
+ public
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State
getState() {
+ return state_;
+ }
+
+ // required .TableName table = 2;
+ public static final int TABLE_FIELD_NUMBER = 2;
+ private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName
table_;
+ /**
+ * <code>required .TableName table = 2;</code>
+ */
+ public boolean hasTable() {
+ return ((bitField0_ & 0x00000002) == 0x00000002);
+ }
+ /**
+ * <code>required .TableName table = 2;</code>
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName
getTable() {
+ return table_;
+ }
+ /**
+ * <code>required .TableName table = 2;</code>
+ */
+ public
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder
getTableOrBuilder() {
+ return table_;
+ }
+
+ // optional uint64 timestamp = 3;
+ public static final int TIMESTAMP_FIELD_NUMBER = 3;
+ private long timestamp_;
+ /**
+ * <code>optional uint64 timestamp = 3;</code>
+ */
+ public boolean hasTimestamp() {
+ return ((bitField0_ & 0x00000004) == 0x00000004);
+ }
+ /**
+ * <code>optional uint64 timestamp = 3;</code>
+ */
+ public long getTimestamp() {
+ return timestamp_;
+ }
+
+ private void initFields() {
+ state_ =
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State.ENABLED;
+ table_ =
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ timestamp_ = 0L;
+ }
+ private byte memoizedIsInitialized = -1;
+ public final boolean isInitialized() {
+ byte isInitialized = memoizedIsInitialized;
+ if (isInitialized != -1) return isInitialized == 1;
+
+ if (!hasState()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
+ if (!hasTable()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
+ if (!getTable().isInitialized()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
+ memoizedIsInitialized = 1;
+ return true;
+ }
+
+ public void writeTo(com.google.protobuf.CodedOutputStream output)
+ throws java.io.IOException {
+ getSerializedSize();
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ output.writeEnum(1, state_.getNumber());
+ }
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ output.writeMessage(2, table_);
+ }
+ if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ output.writeUInt64(3, timestamp_);
+ }
+ getUnknownFields().writeTo(output);
+ }
+
+ private int memoizedSerializedSize = -1;
+ public int getSerializedSize() {
+ int size = memoizedSerializedSize;
+ if (size != -1) return size;
+
+ size = 0;
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeEnumSize(1, state_.getNumber());
+ }
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeMessageSize(2, table_);
+ }
+ if (((bitField0_ & 0x00000004) == 0x00000004)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeUInt64Size(3, timestamp_);
+ }
+ size += getUnknownFields().getSerializedSize();
+ memoizedSerializedSize = size;
+ return size;
+ }
+
+ private static final long serialVersionUID = 0L;
+ @java.lang.Override
+ protected java.lang.Object writeReplace()
+ throws java.io.ObjectStreamException {
+ return super.writeReplace();
+ }
+
+ @java.lang.Override
+ public boolean equals(final java.lang.Object obj) {
+ if (obj == this) {
+ return true;
+ }
+ if (!(obj instanceof
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState)) {
+ return super.equals(obj);
+ }
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState other
= (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState) obj;
+
+ boolean result = true;
+ result = result && (hasState() == other.hasState());
+ if (hasState()) {
+ result = result &&
+ (getState() == other.getState());
+ }
+ result = result && (hasTable() == other.hasTable());
+ if (hasTable()) {
+ result = result && getTable()
+ .equals(other.getTable());
+ }
+ result = result && (hasTimestamp() == other.hasTimestamp());
+ if (hasTimestamp()) {
+ result = result && (getTimestamp()
+ == other.getTimestamp());
+ }
+ result = result &&
+ getUnknownFields().equals(other.getUnknownFields());
+ return result;
+ }
+
+ private int memoizedHashCode = 0;
+ @java.lang.Override
+ public int hashCode() {
+ if (memoizedHashCode != 0) {
+ return memoizedHashCode;
+ }
+ int hash = 41;
+ hash = (19 * hash) + getDescriptorForType().hashCode();
+ if (hasState()) {
+ hash = (37 * hash) + STATE_FIELD_NUMBER;
+ hash = (53 * hash) + hashEnum(getState());
+ }
+ if (hasTable()) {
+ hash = (37 * hash) + TABLE_FIELD_NUMBER;
+ hash = (53 * hash) + getTable().hashCode();
+ }
+ if (hasTimestamp()) {
+ hash = (37 * hash) + TIMESTAMP_FIELD_NUMBER;
+ hash = (53 * hash) + hashLong(getTimestamp());
+ }
+ hash = (29 * hash) + getUnknownFields().hashCode();
+ memoizedHashCode = hash;
+ return hash;
+ }
+
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState parseFrom(
+ com.google.protobuf.ByteString data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState parseFrom(
+ com.google.protobuf.ByteString data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState
parseFrom(byte[] data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState parseFrom(
+ byte[] data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState
parseFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState parseFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState
parseDelimitedFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input);
+ }
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState
parseDelimitedFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ }
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState parseFrom(
+ com.google.protobuf.CodedInputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState parseFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+
+ public static Builder newBuilder() { return Builder.create(); }
+ public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder
newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState
prototype) {
+ return newBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() { return newBuilder(this); }
+
+ @java.lang.Override
+ protected Builder newBuilderForType(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ Builder builder = new Builder(parent);
+ return builder;
+ }
+ /**
+ * Protobuf type {@code TableState}
+ *
+ * <pre>
+ ** Denotes state of the table
+ * </pre>
+ */
+ public static final class Builder extends
+ com.google.protobuf.GeneratedMessage.Builder<Builder>
+ implements
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableStateOrBuilder {
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TableState_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TableState_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.class,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.Builder.class);
+ }
+
+ // Construct using
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.newBuilder()
+ private Builder() {
+ maybeForceBuilderInitialization();
+ }
+
+ private Builder(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ super(parent);
+ maybeForceBuilderInitialization();
+ }
+ private void maybeForceBuilderInitialization() {
+ if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ getTableFieldBuilder();
+ }
+ }
+ private static Builder create() {
+ return new Builder();
+ }
+
+ public Builder clear() {
+ super.clear();
+ state_ =
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State.ENABLED;
+ bitField0_ = (bitField0_ & ~0x00000001);
+ if (tableBuilder_ == null) {
+ table_ =
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ } else {
+ tableBuilder_.clear();
+ }
+ bitField0_ = (bitField0_ & ~0x00000002);
+ timestamp_ = 0L;
+ bitField0_ = (bitField0_ & ~0x00000004);
+ return this;
+ }
+
+ public Builder clone() {
+ return create().mergeFrom(buildPartial());
+ }
+
+ public com.google.protobuf.Descriptors.Descriptor
+ getDescriptorForType() {
+ return
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TableState_descriptor;
+ }
+
+ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState
getDefaultInstanceForType() {
+ return
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.getDefaultInstance();
+ }
+
+ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState
build() {
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState
result = buildPartial();
+ if (!result.isInitialized()) {
+ throw newUninitializedMessageException(result);
+ }
+ return result;
+ }
+
+ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState
buildPartial() {
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState
result = new
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState(this);
+ int from_bitField0_ = bitField0_;
+ int to_bitField0_ = 0;
+ if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+ to_bitField0_ |= 0x00000001;
+ }
+ result.state_ = state_;
+ if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
+ to_bitField0_ |= 0x00000002;
+ }
+ if (tableBuilder_ == null) {
+ result.table_ = table_;
+ } else {
+ result.table_ = tableBuilder_.build();
+ }
+ if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
+ to_bitField0_ |= 0x00000004;
+ }
+ result.timestamp_ = timestamp_;
+ result.bitField0_ = to_bitField0_;
+ onBuilt();
+ return result;
+ }
+
+ public Builder mergeFrom(com.google.protobuf.Message other) {
+ if (other instanceof
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState) {
+ return
mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState)other);
+ } else {
+ super.mergeFrom(other);
+ return this;
+ }
+ }
+
+ public Builder
mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState
other) {
+ if (other ==
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.getDefaultInstance())
return this;
+ if (other.hasState()) {
+ setState(other.getState());
+ }
+ if (other.hasTable()) {
+ mergeTable(other.getTable());
+ }
+ if (other.hasTimestamp()) {
+ setTimestamp(other.getTimestamp());
+ }
+ this.mergeUnknownFields(other.getUnknownFields());
+ return this;
+ }
+
+ public final boolean isInitialized() {
+ if (!hasState()) {
+
+ return false;
+ }
+ if (!hasTable()) {
+
+ return false;
+ }
+ if (!getTable().isInitialized()) {
+
+ return false;
+ }
+ return true;
+ }
+
+ public Builder mergeFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState
parsedMessage = null;
+ try {
+ parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ parsedMessage =
(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState)
e.getUnfinishedMessage();
+ throw e;
+ } finally {
+ if (parsedMessage != null) {
+ mergeFrom(parsedMessage);
+ }
+ }
+ return this;
+ }
+ private int bitField0_;
+
+ // required .TableState.State state = 1;
+ private
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State state_
=
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State.ENABLED;
+ /**
+ * <code>required .TableState.State state = 1;</code>
+ *
+ * <pre>
+ * This is the table's state.
+ * </pre>
+ */
+ public boolean hasState() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * <code>required .TableState.State state = 1;</code>
+ *
+ * <pre>
+ * This is the table's state.
+ * </pre>
+ */
+ public
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State
getState() {
+ return state_;
+ }
+ /**
+ * <code>required .TableState.State state = 1;</code>
+ *
+ * <pre>
+ * This is the table's state.
+ * </pre>
+ */
+ public Builder
setState(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State
value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ bitField0_ |= 0x00000001;
+ state_ = value;
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>required .TableState.State state = 1;</code>
+ *
+ * <pre>
+ * This is the table's state.
+ * </pre>
+ */
+ public Builder clearState() {
+ bitField0_ = (bitField0_ & ~0x00000001);
+ state_ =
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State.ENABLED;
+ onChanged();
+ return this;
+ }
+
+ // required .TableName table = 2;
+ private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName
table_ =
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ private com.google.protobuf.SingleFieldBuilder<
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
tableBuilder_;
+ /**
+ * <code>required .TableName table = 2;</code>
+ */
+ public boolean hasTable() {
+ return ((bitField0_ & 0x00000002) == 0x00000002);
+ }
+ /**
+ * <code>required .TableName table = 2;</code>
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName
getTable() {
+ if (tableBuilder_ == null) {
+ return table_;
+ } else {
+ return tableBuilder_.getMessage();
+ }
+ }
+ /**
+ * <code>required .TableName table = 2;</code>
+ */
+ public Builder
setTable(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName
value) {
+ if (tableBuilder_ == null) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ table_ = value;
+ onChanged();
+ } else {
+ tableBuilder_.setMessage(value);
+ }
+ bitField0_ |= 0x00000002;
+ return this;
+ }
+ /**
+ * <code>required .TableName table = 2;</code>
+ */
+ public Builder setTable(
+
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder
builderForValue) {
+ if (tableBuilder_ == null) {
+ table_ = builderForValue.build();
+ onChanged();
+ } else {
+ tableBuilder_.setMessage(builderForValue.build());
+ }
+ bitField0_ |= 0x00000002;
+ return this;
+ }
+ /**
+ * <code>required .TableName table = 2;</code>
+ */
+ public Builder
mergeTable(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName
value) {
+ if (tableBuilder_ == null) {
+ if (((bitField0_ & 0x00000002) == 0x00000002) &&
+ table_ !=
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance())
{
+ table_ =
+
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(table_).mergeFrom(value).buildPartial();
+ } else {
+ table_ = value;
+ }
+ onChanged();
+ } else {
+ tableBuilder_.mergeFrom(value);
+ }
+ bitField0_ |= 0x00000002;
+ return this;
+ }
+ /**
+ * <code>required .TableName table = 2;</code>
+ */
+ public Builder clearTable() {
+ if (tableBuilder_ == null) {
+ table_ =
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance();
+ onChanged();
+ } else {
+ tableBuilder_.clear();
+ }
+ bitField0_ = (bitField0_ & ~0x00000002);
+ return this;
+ }
+ /**
+ * <code>required .TableName table = 2;</code>
+ */
+ public
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder
getTableBuilder() {
+ bitField0_ |= 0x00000002;
+ onChanged();
+ return getTableFieldBuilder().getBuilder();
+ }
+ /**
+ * <code>required .TableName table = 2;</code>
+ */
+ public
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder
getTableOrBuilder() {
+ if (tableBuilder_ != null) {
+ return tableBuilder_.getMessageOrBuilder();
+ } else {
+ return table_;
+ }
+ }
+ /**
+ * <code>required .TableName table = 2;</code>
+ */
+ private com.google.protobuf.SingleFieldBuilder<
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>
+ getTableFieldBuilder() {
+ if (tableBuilder_ == null) {
+ tableBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
+ table_,
+ getParentForChildren(),
+ isClean());
+ table_ = null;
+ }
+ return tableBuilder_;
+ }
+
+ // optional uint64 timestamp = 3;
+ private long timestamp_ ;
+ /**
+ * <code>optional uint64 timestamp = 3;</code>
+ */
+ public boolean hasTimestamp() {
+ return ((bitField0_ & 0x00000004) == 0x00000004);
+ }
+ /**
+ * <code>optional uint64 timestamp = 3;</code>
+ */
+ public long getTimestamp() {
+ return timestamp_;
+ }
+ /**
+ * <code>optional uint64 timestamp = 3;</code>
+ */
+ public Builder setTimestamp(long value) {
+ bitField0_ |= 0x00000004;
+ timestamp_ = value;
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>optional uint64 timestamp = 3;</code>
+ */
+ public Builder clearTimestamp() {
+ bitField0_ = (bitField0_ & ~0x00000004);
+ timestamp_ = 0L;
+ onChanged();
+ return this;
+ }
+
+ // @@protoc_insertion_point(builder_scope:TableState)
+ }
+
+ static {
+ defaultInstance = new TableState(true);
+ defaultInstance.initFields();
+ }
+
+ // @@protoc_insertion_point(class_scope:TableState)
+ }
+
+ public interface TableDescriptorOrBuilder
+ extends com.google.protobuf.MessageOrBuilder {
+
+ // required .TableSchema schema = 1;
+ /**
+ * <code>required .TableSchema schema = 1;</code>
+ */
+ boolean hasSchema();
+ /**
+ * <code>required .TableSchema schema = 1;</code>
+ */
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema
getSchema();
+ /**
+ * <code>required .TableSchema schema = 1;</code>
+ */
+
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder
getSchemaOrBuilder();
+
+ // optional .TableState.State state = 2 [default = ENABLED];
+ /**
+ * <code>optional .TableState.State state = 2 [default = ENABLED];</code>
+ */
+ boolean hasState();
+ /**
+ * <code>optional .TableState.State state = 2 [default = ENABLED];</code>
+ */
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State
getState();
+ }
+ /**
+ * Protobuf type {@code TableDescriptor}
+ *
+ * <pre>
+ ** On HDFS representation of table state.
+ * </pre>
+ */
+ public static final class TableDescriptor extends
+ com.google.protobuf.GeneratedMessage
+ implements TableDescriptorOrBuilder {
+ // Use TableDescriptor.newBuilder() to construct.
+ private TableDescriptor(com.google.protobuf.GeneratedMessage.Builder<?>
builder) {
+ super(builder);
+ this.unknownFields = builder.getUnknownFields();
+ }
+ private TableDescriptor(boolean noInit) { this.unknownFields =
com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
+ private static final TableDescriptor defaultInstance;
+ public static TableDescriptor getDefaultInstance() {
+ return defaultInstance;
+ }
+
+ public TableDescriptor getDefaultInstanceForType() {
+ return defaultInstance;
+ }
+
+ private final com.google.protobuf.UnknownFieldSet unknownFields;
+ @java.lang.Override
+ public final com.google.protobuf.UnknownFieldSet
+ getUnknownFields() {
+ return this.unknownFields;
+ }
+ private TableDescriptor(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ initFields();
+ int mutable_bitField0_ = 0;
+ com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+ com.google.protobuf.UnknownFieldSet.newBuilder();
+ try {
+ boolean done = false;
+ while (!done) {
+ int tag = input.readTag();
+ switch (tag) {
+ case 0:
+ done = true;
+ break;
+ default: {
+ if (!parseUnknownField(input, unknownFields,
+ extensionRegistry, tag)) {
+ done = true;
+ }
+ break;
+ }
+ case 10: {
+
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder
subBuilder = null;
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ subBuilder = schema_.toBuilder();
+ }
+ schema_ =
input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.PARSER,
extensionRegistry);
+ if (subBuilder != null) {
+ subBuilder.mergeFrom(schema_);
+ schema_ = subBuilder.buildPartial();
+ }
+ bitField0_ |= 0x00000001;
+ break;
+ }
+ case 16: {
+ int rawValue = input.readEnum();
+
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State value =
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State.valueOf(rawValue);
+ if (value == null) {
+ unknownFields.mergeVarintField(2, rawValue);
+ } else {
+ bitField0_ |= 0x00000002;
+ state_ = value;
+ }
+ break;
+ }
+ }
+ }
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ throw e.setUnfinishedMessage(this);
+ } catch (java.io.IOException e) {
+ throw new com.google.protobuf.InvalidProtocolBufferException(
+ e.getMessage()).setUnfinishedMessage(this);
+ } finally {
+ this.unknownFields = unknownFields.build();
+ makeExtensionsImmutable();
+ }
+ }
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TableDescriptor_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TableDescriptor_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor.class,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor.Builder.class);
+ }
+
+ public static com.google.protobuf.Parser<TableDescriptor> PARSER =
+ new com.google.protobuf.AbstractParser<TableDescriptor>() {
+ public TableDescriptor parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new TableDescriptor(input, extensionRegistry);
+ }
+ };
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<TableDescriptor> getParserForType() {
+ return PARSER;
+ }
+
+ private int bitField0_;
+ // required .TableSchema schema = 1;
+ public static final int SCHEMA_FIELD_NUMBER = 1;
+ private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema
schema_;
+ /**
+ * <code>required .TableSchema schema = 1;</code>
+ */
+ public boolean hasSchema() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * <code>required .TableSchema schema = 1;</code>
+ */
+ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema
getSchema() {
+ return schema_;
+ }
+ /**
+ * <code>required .TableSchema schema = 1;</code>
+ */
+ public
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder
getSchemaOrBuilder() {
+ return schema_;
+ }
+
+ // optional .TableState.State state = 2 [default = ENABLED];
+ public static final int STATE_FIELD_NUMBER = 2;
+ private
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State state_;
+ /**
+ * <code>optional .TableState.State state = 2 [default = ENABLED];</code>
+ */
+ public boolean hasState() {
+ return ((bitField0_ & 0x00000002) == 0x00000002);
+ }
+ /**
+ * <code>optional .TableState.State state = 2 [default = ENABLED];</code>
+ */
+ public
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State
getState() {
+ return state_;
+ }
+
+ private void initFields() {
+ schema_ =
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance();
+ state_ =
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State.ENABLED;
+ }
+ private byte memoizedIsInitialized = -1;
+ public final boolean isInitialized() {
+ byte isInitialized = memoizedIsInitialized;
+ if (isInitialized != -1) return isInitialized == 1;
+
+ if (!hasSchema()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
+ if (!getSchema().isInitialized()) {
+ memoizedIsInitialized = 0;
+ return false;
+ }
+ memoizedIsInitialized = 1;
+ return true;
+ }
+
+ public void writeTo(com.google.protobuf.CodedOutputStream output)
+ throws java.io.IOException {
+ getSerializedSize();
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ output.writeMessage(1, schema_);
+ }
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ output.writeEnum(2, state_.getNumber());
+ }
+ getUnknownFields().writeTo(output);
+ }
+
+ private int memoizedSerializedSize = -1;
+ public int getSerializedSize() {
+ int size = memoizedSerializedSize;
+ if (size != -1) return size;
+
+ size = 0;
+ if (((bitField0_ & 0x00000001) == 0x00000001)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeMessageSize(1, schema_);
+ }
+ if (((bitField0_ & 0x00000002) == 0x00000002)) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeEnumSize(2, state_.getNumber());
+ }
+ size += getUnknownFields().getSerializedSize();
+ memoizedSerializedSize = size;
+ return size;
+ }
+
+ private static final long serialVersionUID = 0L;
+ @java.lang.Override
+ protected java.lang.Object writeReplace()
+ throws java.io.ObjectStreamException {
+ return super.writeReplace();
+ }
+
+ @java.lang.Override
+ public boolean equals(final java.lang.Object obj) {
+ if (obj == this) {
+ return true;
+ }
+ if (!(obj instanceof
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor)) {
+ return super.equals(obj);
+ }
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor
other =
(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor) obj;
+
+ boolean result = true;
+ result = result && (hasSchema() == other.hasSchema());
+ if (hasSchema()) {
+ result = result && getSchema()
+ .equals(other.getSchema());
+ }
+ result = result && (hasState() == other.hasState());
+ if (hasState()) {
+ result = result &&
+ (getState() == other.getState());
+ }
+ result = result &&
+ getUnknownFields().equals(other.getUnknownFields());
+ return result;
+ }
+
+ private int memoizedHashCode = 0;
+ @java.lang.Override
+ public int hashCode() {
+ if (memoizedHashCode != 0) {
+ return memoizedHashCode;
+ }
+ int hash = 41;
+ hash = (19 * hash) + getDescriptorForType().hashCode();
+ if (hasSchema()) {
+ hash = (37 * hash) + SCHEMA_FIELD_NUMBER;
+ hash = (53 * hash) + getSchema().hashCode();
+ }
+ if (hasState()) {
+ hash = (37 * hash) + STATE_FIELD_NUMBER;
+ hash = (53 * hash) + hashEnum(getState());
+ }
+ hash = (29 * hash) + getUnknownFields().hashCode();
+ memoizedHashCode = hash;
+ return hash;
+ }
+
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor
parseFrom(
+ com.google.protobuf.ByteString data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor
parseFrom(
+ com.google.protobuf.ByteString data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor
parseFrom(byte[] data)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data);
+ }
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor
parseFrom(
+ byte[] data,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return PARSER.parseFrom(data, extensionRegistry);
+ }
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor
parseFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor
parseFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor
parseDelimitedFrom(java.io.InputStream input)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input);
+ }
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor
parseDelimitedFrom(
+ java.io.InputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ }
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor
parseFrom(
+ com.google.protobuf.CodedInputStream input)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input);
+ }
+ public static
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor
parseFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ return PARSER.parseFrom(input, extensionRegistry);
+ }
+
+ public static Builder newBuilder() { return Builder.create(); }
+ public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder
newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor
prototype) {
+ return newBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() { return newBuilder(this); }
+
+ @java.lang.Override
+ protected Builder newBuilderForType(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ Builder builder = new Builder(parent);
+ return builder;
+ }
+ /**
+ * Protobuf type {@code TableDescriptor}
+ *
+ * <pre>
+ ** On HDFS representation of table state.
+ * </pre>
+ */
+ public static final class Builder extends
+ com.google.protobuf.GeneratedMessage.Builder<Builder>
+ implements
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptorOrBuilder
{
+ public static final com.google.protobuf.Descriptors.Descriptor
+ getDescriptor() {
+ return
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TableDescriptor_descriptor;
+ }
+
+ protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internalGetFieldAccessorTable() {
+ return
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TableDescriptor_fieldAccessorTable
+ .ensureFieldAccessorsInitialized(
+
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor.class,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor.Builder.class);
+ }
+
+ // Construct using
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor.newBuilder()
+ private Builder() {
+ maybeForceBuilderInitialization();
+ }
+
+ private Builder(
+ com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ super(parent);
+ maybeForceBuilderInitialization();
+ }
+ private void maybeForceBuilderInitialization() {
+ if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ getSchemaFieldBuilder();
+ }
+ }
+ private static Builder create() {
+ return new Builder();
+ }
+
+ public Builder clear() {
+ super.clear();
+ if (schemaBuilder_ == null) {
+ schema_ =
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance();
+ } else {
+ schemaBuilder_.clear();
+ }
+ bitField0_ = (bitField0_ & ~0x00000001);
+ state_ =
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State.ENABLED;
+ bitField0_ = (bitField0_ & ~0x00000002);
+ return this;
+ }
+
+ public Builder clone() {
+ return create().mergeFrom(buildPartial());
+ }
+
+ public com.google.protobuf.Descriptors.Descriptor
+ getDescriptorForType() {
+ return
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_TableDescriptor_descriptor;
+ }
+
+ public
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor
getDefaultInstanceForType() {
+ return
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor.getDefaultInstance();
+ }
+
+ public
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor build() {
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor
result = buildPartial();
+ if (!result.isInitialized()) {
+ throw newUninitializedMessageException(result);
+ }
+ return result;
+ }
+
+ public
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor
buildPartial() {
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor
result = new
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor(this);
+ int from_bitField0_ = bitField0_;
+ int to_bitField0_ = 0;
+ if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+ to_bitField0_ |= 0x00000001;
+ }
+ if (schemaBuilder_ == null) {
+ result.schema_ = schema_;
+ } else {
+ result.schema_ = schemaBuilder_.build();
+ }
+ if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
+ to_bitField0_ |= 0x00000002;
+ }
+ result.state_ = state_;
+ result.bitField0_ = to_bitField0_;
+ onBuilt();
+ return result;
+ }
+
+ public Builder mergeFrom(com.google.protobuf.Message other) {
+ if (other instanceof
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor) {
+ return
mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor)other);
+ } else {
+ super.mergeFrom(other);
+ return this;
+ }
+ }
+
+ public Builder
mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor
other) {
+ if (other ==
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor.getDefaultInstance())
return this;
+ if (other.hasSchema()) {
+ mergeSchema(other.getSchema());
+ }
+ if (other.hasState()) {
+ setState(other.getState());
+ }
+ this.mergeUnknownFields(other.getUnknownFields());
+ return this;
+ }
+
+ public final boolean isInitialized() {
+ if (!hasSchema()) {
+
+ return false;
+ }
+ if (!getSchema().isInitialized()) {
+
+ return false;
+ }
+ return true;
+ }
+
+ public Builder mergeFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws java.io.IOException {
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor
parsedMessage = null;
+ try {
+ parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+ } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+ parsedMessage =
(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableDescriptor)
e.getUnfinishedMessage();
+ throw e;
+ } finally {
+ if (parsedMessage != null) {
+ mergeFrom(parsedMessage);
+ }
+ }
+ return this;
+ }
+ private int bitField0_;
+
+ // required .TableSchema schema = 1;
+ private
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema schema_ =
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance();
+ private com.google.protobuf.SingleFieldBuilder<
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder>
schemaBuilder_;
+ /**
+ * <code>required .TableSchema schema = 1;</code>
+ */
+ public boolean hasSchema() {
+ return ((bitField0_ & 0x00000001) == 0x00000001);
+ }
+ /**
+ * <code>required .TableSchema schema = 1;</code>
+ */
+ public
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema getSchema() {
+ if (schemaBuilder_ == null) {
+ return schema_;
+ } else {
+ return schemaBuilder_.getMessage();
+ }
+ }
+ /**
+ * <code>required .TableSchema schema = 1;</code>
+ */
+ public Builder
setSchema(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema
value) {
+ if (schemaBuilder_ == null) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ schema_ = value;
+ onChanged();
+ } else {
+ schemaBuilder_.setMessage(value);
+ }
+ bitField0_ |= 0x00000001;
+ return this;
+ }
+ /**
+ * <code>required .TableSchema schema = 1;</code>
+ */
+ public Builder setSchema(
+
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder
builderForValue) {
+ if (schemaBuilder_ == null) {
+ schema_ = builderForValue.build();
+ onChanged();
+ } else {
+ schemaBuilder_.setMessage(builderForValue.build());
+ }
+ bitField0_ |= 0x00000001;
+ return this;
+ }
+ /**
+ * <code>required .TableSchema schema = 1;</code>
+ */
+ public Builder
mergeSchema(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema
value) {
+ if (schemaBuilder_ == null) {
+ if (((bitField0_ & 0x00000001) == 0x00000001) &&
+ schema_ !=
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance())
{
+ schema_ =
+
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.newBuilder(schema_).mergeFrom(value).buildPartial();
+ } else {
+ schema_ = value;
+ }
+ onChanged();
+ } else {
+ schemaBuilder_.mergeFrom(value);
+ }
+ bitField0_ |= 0x00000001;
+ return this;
+ }
+ /**
+ * <code>required .TableSchema schema = 1;</code>
+ */
+ public Builder clearSchema() {
+ if (schemaBuilder_ == null) {
+ schema_ =
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance();
+ onChanged();
+ } else {
+ schemaBuilder_.clear();
+ }
+ bitField0_ = (bitField0_ & ~0x00000001);
+ return this;
+ }
+ /**
+ * <code>required .TableSchema schema = 1;</code>
+ */
+ public
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder
getSchemaBuilder() {
+ bitField0_ |= 0x00000001;
+ onChanged();
+ return getSchemaFieldBuilder().getBuilder();
+ }
+ /**
+ * <code>required .TableSchema schema = 1;</code>
+ */
+ public
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder
getSchemaOrBuilder() {
+ if (schemaBuilder_ != null) {
+ return schemaBuilder_.getMessageOrBuilder();
+ } else {
+ return schema_;
+ }
+ }
+ /**
+ * <code>required .TableSchema schema = 1;</code>
+ */
+ private com.google.protobuf.SingleFieldBuilder<
+ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder>
+ getSchemaFieldBuilder() {
+ if (schemaBuilder_ == null) {
+ schemaBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchema.Builder,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableSchemaOrBuilder>(
+ schema_,
+ getParentForChildren(),
+ isClean());
+ schema_ = null;
+ }
+ return schemaBuilder_;
+ }
+
+ // optional .TableState.State state = 2 [default = ENABLED];
+ private
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State state_
=
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State.ENABLED;
+ /**
+ * <code>optional .TableState.State state = 2 [default = ENABLED];</code>
+ */
+ public boolean hasState() {
+ return ((bitField0_ & 0x00000002) == 0x00000002);
+ }
+ /**
+ * <code>optional .TableState.State state = 2 [default = ENABLED];</code>
+ */
+ public
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State
getState() {
+ return state_;
+ }
+ /**
+ * <code>optional .TableState.State state = 2 [default = ENABLED];</code>
+ */
+ public Builder
setState(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State
value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ bitField0_ |= 0x00000002;
+ state_ = value;
+ onChanged();
+ return this;
+ }
+ /**
+ * <code>optional .TableState.State state = 2 [default = ENABLED];</code>
+ */
+ public Builder clearState() {
+ bitField0_ = (bitField0_ & ~0x00000002);
+ state_ =
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableState.State.ENABLED;
+ onChanged();
+ return this;
+ }
+
+ // @@protoc_insertion_point(builder_scope:TableDescriptor)
+ }
+
+ static {
+ defaultInstance = new TableDescriptor(true);
+ defaultInstance.initFields();
+ }
+
+ // @@protoc_insertion_point(class_scope:TableDescriptor)
+ }
+
public interface ColumnFamilySchemaOrBuilder
extends com.google.protobuf.MessageOrBuilder {
@@ -16372,6 +17942,16 @@ public final class HBaseProtos {
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_TableSchema_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
+ internal_static_TableState_descriptor;
+ private static
+ com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internal_static_TableState_fieldAccessorTable;
+ private static com.google.protobuf.Descriptors.Descriptor
+ internal_static_TableDescriptor_descriptor;
+ private static
+ com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ internal_static_TableDescriptor_fieldAccessorTable;
+ private static com.google.protobuf.Descriptors.Descriptor
internal_static_ColumnFamilySchema_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
@@ -16486,47 +18066,53 @@ public final class HBaseProtos {
"Name\022#\n\nattributes\030\002 \003(\0132\017.BytesBytesPai" +
"r\022,\n\017column_families\030\003 \003(\0132\023.ColumnFamil" +
"ySchema\022&\n\rconfiguration\030\004 \003(\0132\017.NameStr" +
- "ingPair\"o\n\022ColumnFamilySchema\022\014\n\004name\030\001 " +
- "\002(\014\022#\n\nattributes\030\002 \003(\0132\017.BytesBytesPair" +
- "\022&\n\rconfiguration\030\003 \003(\0132\017.NameStringPair" +
- "\"\232\001\n\nRegionInfo\022\021\n\tregion_id\030\001
\002(\004\022\036\n\nta",
- "ble_name\030\002 \002(\0132\n.TableName\022\021\n\tstart_key\030" +
- "\003 \001(\014\022\017\n\007end_key\030\004
\001(\014\022\017\n\007offline\030\005 \001(\010\022" +
- "\r\n\005split\030\006 \001(\010\022\025\n\nreplica_id\030\007
\001(\005:\0010\"1\n" +
- "\014FavoredNodes\022!\n\014favored_node\030\001 \003(\0132\013.Se" +
- "rverName\"\225\001\n\017RegionSpecifier\0222\n\004type\030\001 \002" +
- "(\0162$.RegionSpecifier.RegionSpecifierType" +
- "\022\r\n\005value\030\002
\002(\014\"?\n\023RegionSpecifierType\022\017" +
- "\n\013REGION_NAME\020\001\022\027\n\023ENCODED_REGION_NAME\020\002" +
- "\"%\n\tTimeRange\022\014\n\004from\030\001
\001(\004\022\n\n\002to\030\002 \001(\004\"" +
- "A\n\nServerName\022\021\n\thost_name\030\001 \002(\t\022\014\n\004port",
- "\030\002 \001(\r\022\022\n\nstart_code\030\003
\001(\004\"\033\n\013Coprocesso" +
- "r\022\014\n\004name\030\001
\002(\t\"-\n\016NameStringPair\022\014\n\004nam" +
- "e\030\001 \002(\t\022\r\n\005value\030\002
\002(\t\",\n\rNameBytesPair\022" +
- "\014\n\004name\030\001 \002(\t\022\r\n\005value\030\002
\001(\014\"/\n\016BytesByt" +
- "esPair\022\r\n\005first\030\001 \002(\014\022\016\n\006second\030\002
\002(\014\",\n" +
- "\rNameInt64Pair\022\014\n\004name\030\001
\001(\t\022\r\n\005value\030\002 " +
- "\001(\003\"\314\001\n\023SnapshotDescription\022\014\n\004name\030\001
\002(" +
- "\t\022\r\n\005table\030\002 \001(\t\022\030\n\rcreation_time\030\003
\001(\003:" +
- "\0010\022.\n\004type\030\004 \001(\0162\031.SnapshotDescription.T" +
- "ype:\005FLUSH\022\017\n\007version\030\005
\001(\005\022\r\n\005owner\030\006 \001",
-
"(\t\".\n\004Type\022\014\n\010DISABLED\020\000\022\t\n\005FLUSH\020\001\022\r\n\tS"
+
- "KIPFLUSH\020\002\"}\n\024ProcedureDescription\022\021\n\tsi" +
- "gnature\030\001 \002(\t\022\020\n\010instance\030\002
\001(\t\022\030\n\rcreat" +
- "ion_time\030\003 \001(\003:\0010\022&\n\rconfiguration\030\004
\003(\013" +
- "2\017.NameStringPair\"\n\n\010EmptyMsg\"\033\n\007LongMsg" +
- "\022\020\n\010long_msg\030\001
\002(\003\"\037\n\tDoubleMsg\022\022\n\ndoubl" +
- "e_msg\030\001 \002(\001\"\'\n\rBigDecimalMsg\022\026\n\016bigdecim" +
- "al_msg\030\001
\002(\014\"5\n\004UUID\022\026\n\016least_sig_bits\030\001" +
- " \002(\004\022\025\n\rmost_sig_bits\030\002
\002(\004\"K\n\023Namespace" +
- "Descriptor\022\014\n\004name\030\001 \002(\014\022&\n\rconfiguratio",
- "n\030\002 \003(\0132\017.NameStringPair\"$\n\020RegionServer" +
- "Info\022\020\n\010infoPort\030\001
\001(\005*r\n\013CompareType\022\010\n" +
-
"\004LESS\020\000\022\021\n\rLESS_OR_EQUAL\020\001\022\t\n\005EQUAL\020\002\022\r\n"
+
-
"\tNOT_EQUAL\020\003\022\024\n\020GREATER_OR_EQUAL\020\004\022\013\n\007GR" +
- "EATER\020\005\022\t\n\005NO_OP\020\006B>\n*org.apache.hadoop." +
- "hbase.protobuf.generatedB\013HBaseProtosH\001\240" +
- "\001\001"
+ "ingPair\"\235\001\n\nTableState\022 \n\005state\030\001
\002(\0162\021." +
+ "TableState.State\022\031\n\005table\030\002 \002(\0132\n.TableN" +
+ "ame\022\021\n\ttimestamp\030\003
\001(\004\"?\n\005State\022\013\n\007ENABL" +
+
"ED\020\000\022\014\n\010DISABLED\020\001\022\r\n\tDISABLING\020\002\022\014\n\010ENA",
+ "BLING\020\003\"Z\n\017TableDescriptor\022\034\n\006schema\030\001 \002"
+
+ "(\0132\014.TableSchema\022)\n\005state\030\002 \001(\0162\021.TableS" +
+ "tate.State:\007ENABLED\"o\n\022ColumnFamilySchem" +
+ "a\022\014\n\004name\030\001 \002(\014\022#\n\nattributes\030\002
\003(\0132\017.By" +
+ "tesBytesPair\022&\n\rconfiguration\030\003 \003(\0132\017.Na" +
+ "meStringPair\"\232\001\n\nRegionInfo\022\021\n\tregion_id" +
+ "\030\001 \002(\004\022\036\n\ntable_name\030\002
\002(\0132\n.TableName\022\021" +
+ "\n\tstart_key\030\003 \001(\014\022\017\n\007end_key\030\004
\001(\014\022\017\n\007of" +
+ "fline\030\005 \001(\010\022\r\n\005split\030\006
\001(\010\022\025\n\nreplica_id" +
+ "\030\007 \001(\005:\0010\"1\n\014FavoredNodes\022!\n\014favored_nod",
+ "e\030\001 \003(\0132\013.ServerName\"\225\001\n\017RegionSpecifier" +
+ "\0222\n\004type\030\001 \002(\0162$.RegionSpecifier.RegionS" +
+ "pecifierType\022\r\n\005value\030\002 \002(\014\"?\n\023RegionSpe" +
+ "cifierType\022\017\n\013REGION_NAME\020\001\022\027\n\023ENCODED_R" +
+ "EGION_NAME\020\002\"%\n\tTimeRange\022\014\n\004from\030\001
\001(\004\022" +
+ "\n\n\002to\030\002
\001(\004\"A\n\nServerName\022\021\n\thost_name\030\001" +
+ " \002(\t\022\014\n\004port\030\002
\001(\r\022\022\n\nstart_code\030\003 \001(\004\"\033" +
+ "\n\013Coprocessor\022\014\n\004name\030\001 \002(\t\"-\n\016NameStrin" +
+ "gPair\022\014\n\004name\030\001 \002(\t\022\r\n\005value\030\002
\002(\t\",\n\rNa" +
+ "meBytesPair\022\014\n\004name\030\001 \002(\t\022\r\n\005value\030\002
\001(\014",
+ "\"/\n\016BytesBytesPair\022\r\n\005first\030\001
\002(\014\022\016\n\006sec" +
+ "ond\030\002 \002(\014\",\n\rNameInt64Pair\022\014\n\004name\030\001
\001(\t" +
+ "\022\r\n\005value\030\002
\001(\003\"\314\001\n\023SnapshotDescription\022" +
+ "\014\n\004name\030\001 \002(\t\022\r\n\005table\030\002
\001(\t\022\030\n\rcreation" +
+ "_time\030\003 \001(\003:\0010\022.\n\004type\030\004
\001(\0162\031.SnapshotD" +
+ "escription.Type:\005FLUSH\022\017\n\007version\030\005 \001(\005\022" +
+ "\r\n\005owner\030\006
\001(\t\".\n\004Type\022\014\n\010DISABLED\020\000\022\t\n\005" +
+ "FLUSH\020\001\022\r\n\tSKIPFLUSH\020\002\"}\n\024ProcedureDescr" +
+ "iption\022\021\n\tsignature\030\001
\002(\t\022\020\n\010instance\030\002 " +
+ "\001(\t\022\030\n\rcreation_time\030\003
\001(\003:\0010\022&\n\rconfigu",
+ "ration\030\004 \003(\0132\017.NameStringPair\"\n\n\010EmptyMs" +
+ "g\"\033\n\007LongMsg\022\020\n\010long_msg\030\001
\002(\003\"\037\n\tDouble" +
+ "Msg\022\022\n\ndouble_msg\030\001 \002(\001\"\'\n\rBigDecimalMsg" +
+ "\022\026\n\016bigdecimal_msg\030\001
\002(\014\"5\n\004UUID\022\026\n\016leas" +
+ "t_sig_bits\030\001 \002(\004\022\025\n\rmost_sig_bits\030\002
\002(\004\"" +
+ "K\n\023NamespaceDescriptor\022\014\n\004name\030\001
\002(\014\022&\n\r" +
+ "configuration\030\002 \003(\0132\017.NameStringPair\"$\n\020" +
+ "RegionServerInfo\022\020\n\010infoPort\030\001 \001(\005*r\n\013Co" +
+
"mpareType\022\010\n\004LESS\020\000\022\021\n\rLESS_OR_EQUAL\020\001\022\t" +
+
"\n\005EQUAL\020\002\022\r\n\tNOT_EQUAL\020\003\022\024\n\020GREATER_OR_E",
+
"QUAL\020\004\022\013\n\007GREATER\020\005\022\t\n\005NO_OP\020\006B>\n*org.ap"
+
+ "ache.hadoop.hbase.protobuf.generatedB\013HB" +
+ "aseProtosH\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner
assigner =
new
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
@@ -16545,122 +18131,134 @@ public final class HBaseProtos {
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_TableSchema_descriptor,
new java.lang.String[] { "TableName", "Attributes",
"ColumnFamilies", "Configuration", });
- internal_static_ColumnFamilySchema_descriptor =
+ internal_static_TableState_descriptor =
getDescriptor().getMessageTypes().get(2);
+ internal_static_TableState_fieldAccessorTable = new
+ com.google.protobuf.GeneratedMessage.FieldAccessorTable(
+ internal_static_TableState_descriptor,
+ new java.lang.String[] { "State", "Table", "Timestamp", });
+ internal_static_TableDescriptor_descriptor =
+ getDescriptor().getMessageTypes().get(3);
+ internal_static_TableDescriptor_fieldAccessorTable = new
+ com.google.protobuf.GeneratedMessage.FieldAccessorTable(
+ internal_static_TableDescriptor_descriptor,
+ new java.lang.String[] { "Schema", "State", });
+ internal_static_ColumnFamilySchema_descriptor =
+ getDescriptor().getMessageTypes().get(4);
internal_static_ColumnFamilySchema_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_ColumnFamilySchema_descriptor,
new java.lang.String[] { "Name", "Attributes", "Configuration",
});
internal_static_RegionInfo_descriptor =
- getDescriptor().getMessageTypes().get(3);
+ getDescriptor().getMessageTypes().get(5);
internal_static_RegionInfo_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_RegionInfo_descriptor,
new java.lang.String[] { "RegionId", "TableName", "StartKey",
"EndKey", "Offline", "Split", "ReplicaId", });
internal_static_FavoredNodes_descriptor =
- getDescriptor().getMessageTypes().get(4);
+ getDescriptor().getMessageTypes().get(6);
internal_static_FavoredNodes_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_FavoredNodes_descriptor,
new java.lang.String[] { "FavoredNode", });
internal_static_RegionSpecifier_descriptor =
- getDescriptor().getMessageTypes().get(5);
+ getDescriptor().getMessageTypes().get(7);
internal_static_RegionSpecifier_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_RegionSpecifier_descriptor,
new java.lang.String[] { "Type", "Value", });
internal_static_TimeRange_descriptor =
- getDescriptor().getMessageTypes().get(6);
+ getDescriptor().getMessageTypes().get(8);
internal_static_TimeRange_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_TimeRange_descriptor,
new java.lang.String[] { "From", "To", });
internal_static_ServerName_descriptor =
- getDescriptor().getMessageTypes().get(7);
+ getDescriptor().getMessageTypes().get(9);
internal_static_ServerName_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_ServerName_descriptor,
new java.lang.String[] { "HostName", "Port", "StartCode", });
internal_static_Coprocessor_descriptor =
- getDescriptor().getMessageTypes().get(8);
+ getDescriptor().getMessageTypes().get(10);
internal_static_Coprocessor_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_Coprocessor_descriptor,
new java.lang.String[] { "Name", });
internal_static_NameStringPair_descriptor =
- getDescriptor().getMessageTypes().get(9);
+ getDescriptor().getMessageTypes().get(11);
internal_static_NameStringPair_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_NameStringPair_descriptor,
new java.lang.String[] { "Name", "Value", });
internal_static_NameBytesPair_descriptor =
- getDescriptor().getMessageTypes().get(10);
+ getDescriptor().getMessageTypes().get(12);
internal_static_NameBytesPair_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_NameBytesPair_descriptor,
new java.lang.String[] { "Name", "Value", });
internal_static_BytesBytesPair_descriptor =
- getDescriptor().getMessageTypes().get(11);
+ getDescriptor().getMessageTypes().get(13);
internal_static_BytesBytesPair_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_BytesBytesPair_descriptor,
new java.lang.String[] { "First", "Second", });
internal_static_NameInt64Pair_descriptor =
- getDescriptor().getMessageTypes().get(12);
+ getDescriptor().getMessageTypes().get(14);
internal_static_NameInt64Pair_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_NameInt64Pair_descriptor,
new java.lang.String[] { "Name", "Value", });
internal_static_SnapshotDescription_descriptor =
- getDescriptor().getMessageTypes().get(13);
+ getDescriptor().getMessageTypes().get(15);
internal_static_SnapshotDescription_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_SnapshotDescription_descriptor,
new java.lang.String[] { "Name", "Table", "CreationTime",
"Type", "Version", "Owner", });
internal_static_ProcedureDescription_descriptor =
- getDescriptor().getMessageTypes().get(14);
+ getDescriptor().getMessageTypes().get(16);
internal_static_ProcedureDescription_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_ProcedureDescription_descriptor,
new java.lang.String[] { "Signature", "Instance",
"CreationTime", "Configuration", });
internal_static_EmptyMsg_descriptor =
- getDescriptor().getMessageTypes().get(15);
+ getDescriptor().getMessageTypes().get(17);
internal_static_EmptyMsg_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_EmptyMsg_descriptor,
new java.lang.String[] { });
internal_static_LongMsg_descriptor =
- getDescriptor().getMessageTypes().get(16);
+ getDescriptor().getMessageTypes().get(18);
internal_static_LongMsg_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_LongMsg_descriptor,
new java.lang.String[] { "LongMsg", });
internal_static_DoubleMsg_descriptor =
- getDescriptor().getMessageTypes().get(17);
+ getDescriptor().getMessageTypes().get(19);
internal_static_DoubleMsg_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_DoubleMsg_descriptor,
new java.lang.String[] { "DoubleMsg", });
internal_static_BigDecimalMsg_descriptor =
- getDescriptor().getMessageTypes().get(18);
+ getDescriptor().getMessageTypes().get(20);
internal_static_BigDecimalMsg_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_BigDecimalMsg_descriptor,
new java.lang.String[] { "BigdecimalMsg", });
internal_static_UUID_descriptor =
- getDescriptor().getMessageTypes().get(19);
+ getDescriptor().getMessageTypes().get(21);
internal_static_UUID_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_UUID_descriptor,
new java.lang.String[] { "LeastSigBits", "MostSigBits", });
internal_static_NamespaceDescriptor_descriptor =
- getDescriptor().getMessageTypes().get(20);
+ getDescriptor().getMessageTypes().get(22);
internal_static_NamespaceDescriptor_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_NamespaceDescriptor_descriptor,
new java.lang.String[] { "Name", "Configuration", });
internal_static_RegionServerInfo_descriptor =
- getDescriptor().getMessageTypes().get(21);
+ getDescriptor().getMessageTypes().get(23);
internal_static_RegionServerInfo_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_RegionServerInfo_descriptor,