http://git-wip-us.apache.org/repos/asf/hive/blob/cae3ec16/llap-server/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java ---------------------------------------------------------------------- diff --git a/llap-server/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java b/llap-server/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java index 1500635..8748151 100644 --- a/llap-server/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java +++ b/llap-server/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java @@ -9622,6 +9622,2555 @@ public final class LlapDaemonProtocolProtos { // @@protoc_insertion_point(class_scope:SourceStateUpdatedResponseProto) } + public interface QueryCompleteRequestProtoOrBuilder + extends com.google.protobuf.MessageOrBuilder { + + // optional string query_id = 1; + /** + * <code>optional string query_id = 1;</code> + */ + boolean hasQueryId(); + /** + * <code>optional string query_id = 1;</code> + */ + java.lang.String getQueryId(); + /** + * <code>optional string query_id = 1;</code> + */ + com.google.protobuf.ByteString + getQueryIdBytes(); + + // optional string dag_name = 2; + /** + * <code>optional string dag_name = 2;</code> + */ + boolean hasDagName(); + /** + * <code>optional string dag_name = 2;</code> + */ + java.lang.String getDagName(); + /** + * <code>optional string dag_name = 2;</code> + */ + com.google.protobuf.ByteString + getDagNameBytes(); + + // optional int64 delete_delay = 3 [default = 0]; + /** + * <code>optional int64 delete_delay = 3 [default = 0];</code> + */ + boolean hasDeleteDelay(); + /** + * <code>optional int64 delete_delay = 3 [default = 0];</code> + */ + long getDeleteDelay(); + } + /** + * Protobuf type {@code QueryCompleteRequestProto} + */ + public static final class QueryCompleteRequestProto extends + com.google.protobuf.GeneratedMessage + implements QueryCompleteRequestProtoOrBuilder { + // Use QueryCompleteRequestProto.newBuilder() to construct. + private QueryCompleteRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + super(builder); + this.unknownFields = builder.getUnknownFields(); + } + private QueryCompleteRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + + private static final QueryCompleteRequestProto defaultInstance; + public static QueryCompleteRequestProto getDefaultInstance() { + return defaultInstance; + } + + public QueryCompleteRequestProto getDefaultInstanceForType() { + return defaultInstance; + } + + private final com.google.protobuf.UnknownFieldSet unknownFields; + @java.lang.Override + public final com.google.protobuf.UnknownFieldSet + getUnknownFields() { + return this.unknownFields; + } + private QueryCompleteRequestProto( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + initFields(); + int mutable_bitField0_ = 0; + com.google.protobuf.UnknownFieldSet.Builder unknownFields = + com.google.protobuf.UnknownFieldSet.newBuilder(); + try { + boolean done = false; + while (!done) { + int tag = input.readTag(); + switch (tag) { + case 0: + done = true; + break; + default: { + if (!parseUnknownField(input, unknownFields, + extensionRegistry, tag)) { + done = true; + } + break; + } + case 10: { + bitField0_ |= 0x00000001; + queryId_ = input.readBytes(); + break; + } + case 18: { + bitField0_ |= 0x00000002; + dagName_ = input.readBytes(); + break; + } + case 24: { + bitField0_ |= 0x00000004; + deleteDelay_ = input.readInt64(); + break; + } + } + } + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + throw e.setUnfinishedMessage(this); + } catch (java.io.IOException e) { + throw new com.google.protobuf.InvalidProtocolBufferException( + e.getMessage()).setUnfinishedMessage(this); + } finally { + this.unknownFields = unknownFields.build(); + makeExtensionsImmutable(); + } + } + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteRequestProto_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteRequestProto_fieldAccessorTable + .ensureFieldAccessorsInitialized( + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.Builder.class); + } + + public static com.google.protobuf.Parser<QueryCompleteRequestProto> PARSER = + new com.google.protobuf.AbstractParser<QueryCompleteRequestProto>() { + public QueryCompleteRequestProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new QueryCompleteRequestProto(input, extensionRegistry); + } + }; + + @java.lang.Override + public com.google.protobuf.Parser<QueryCompleteRequestProto> getParserForType() { + return PARSER; + } + + private int bitField0_; + // optional string query_id = 1; + public static final int QUERY_ID_FIELD_NUMBER = 1; + private java.lang.Object queryId_; + /** + * <code>optional string query_id = 1;</code> + */ + public boolean hasQueryId() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + /** + * <code>optional string query_id = 1;</code> + */ + public java.lang.String getQueryId() { + java.lang.Object ref = queryId_; + if (ref instanceof java.lang.String) { + return (java.lang.String) ref; + } else { + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + queryId_ = s; + } + return s; + } + } + /** + * <code>optional string query_id = 1;</code> + */ + public com.google.protobuf.ByteString + getQueryIdBytes() { + java.lang.Object ref = queryId_; + if (ref instanceof java.lang.String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8( + (java.lang.String) ref); + queryId_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + + // optional string dag_name = 2; + public static final int DAG_NAME_FIELD_NUMBER = 2; + private java.lang.Object dagName_; + /** + * <code>optional string dag_name = 2;</code> + */ + public boolean hasDagName() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + /** + * <code>optional string dag_name = 2;</code> + */ + public java.lang.String getDagName() { + java.lang.Object ref = dagName_; + if (ref instanceof java.lang.String) { + return (java.lang.String) ref; + } else { + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + dagName_ = s; + } + return s; + } + } + /** + * <code>optional string dag_name = 2;</code> + */ + public com.google.protobuf.ByteString + getDagNameBytes() { + java.lang.Object ref = dagName_; + if (ref instanceof java.lang.String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8( + (java.lang.String) ref); + dagName_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + + // optional int64 delete_delay = 3 [default = 0]; + public static final int DELETE_DELAY_FIELD_NUMBER = 3; + private long deleteDelay_; + /** + * <code>optional int64 delete_delay = 3 [default = 0];</code> + */ + public boolean hasDeleteDelay() { + return ((bitField0_ & 0x00000004) == 0x00000004); + } + /** + * <code>optional int64 delete_delay = 3 [default = 0];</code> + */ + public long getDeleteDelay() { + return deleteDelay_; + } + + private void initFields() { + queryId_ = ""; + dagName_ = ""; + deleteDelay_ = 0L; + } + private byte memoizedIsInitialized = -1; + public final boolean isInitialized() { + byte isInitialized = memoizedIsInitialized; + if (isInitialized != -1) return isInitialized == 1; + + memoizedIsInitialized = 1; + return true; + } + + public void writeTo(com.google.protobuf.CodedOutputStream output) + throws java.io.IOException { + getSerializedSize(); + if (((bitField0_ & 0x00000001) == 0x00000001)) { + output.writeBytes(1, getQueryIdBytes()); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + output.writeBytes(2, getDagNameBytes()); + } + if (((bitField0_ & 0x00000004) == 0x00000004)) { + output.writeInt64(3, deleteDelay_); + } + getUnknownFields().writeTo(output); + } + + private int memoizedSerializedSize = -1; + public int getSerializedSize() { + int size = memoizedSerializedSize; + if (size != -1) return size; + + size = 0; + if (((bitField0_ & 0x00000001) == 0x00000001)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(1, getQueryIdBytes()); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(2, getDagNameBytes()); + } + if (((bitField0_ & 0x00000004) == 0x00000004)) { + size += com.google.protobuf.CodedOutputStream + .computeInt64Size(3, deleteDelay_); + } + size += getUnknownFields().getSerializedSize(); + memoizedSerializedSize = size; + return size; + } + + private static final long serialVersionUID = 0L; + @java.lang.Override + protected java.lang.Object writeReplace() + throws java.io.ObjectStreamException { + return super.writeReplace(); + } + + @java.lang.Override + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto)) { + return super.equals(obj); + } + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto) obj; + + boolean result = true; + result = result && (hasQueryId() == other.hasQueryId()); + if (hasQueryId()) { + result = result && getQueryId() + .equals(other.getQueryId()); + } + result = result && (hasDagName() == other.hasDagName()); + if (hasDagName()) { + result = result && getDagName() + .equals(other.getDagName()); + } + result = result && (hasDeleteDelay() == other.hasDeleteDelay()); + if (hasDeleteDelay()) { + result = result && (getDeleteDelay() + == other.getDeleteDelay()); + } + result = result && + getUnknownFields().equals(other.getUnknownFields()); + return result; + } + + private int memoizedHashCode = 0; + @java.lang.Override + public int hashCode() { + if (memoizedHashCode != 0) { + return memoizedHashCode; + } + int hash = 41; + hash = (19 * hash) + getDescriptorForType().hashCode(); + if (hasQueryId()) { + hash = (37 * hash) + QUERY_ID_FIELD_NUMBER; + hash = (53 * hash) + getQueryId().hashCode(); + } + if (hasDagName()) { + hash = (37 * hash) + DAG_NAME_FIELD_NUMBER; + hash = (53 * hash) + getDagName().hashCode(); + } + if (hasDeleteDelay()) { + hash = (37 * hash) + DELETE_DELAY_FIELD_NUMBER; + hash = (53 * hash) + hashLong(getDeleteDelay()); + } + hash = (29 * hash) + getUnknownFields().hashCode(); + memoizedHashCode = hash; + return hash; + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom(byte[] data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom( + byte[] data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom(java.io.InputStream input) + throws java.io.IOException { + return PARSER.parseFrom(input); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseFrom(input, extensionRegistry); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseDelimitedFrom(java.io.InputStream input) + throws java.io.IOException { + return PARSER.parseDelimitedFrom(input); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseDelimitedFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseDelimitedFrom(input, extensionRegistry); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom( + com.google.protobuf.CodedInputStream input) + throws java.io.IOException { + return PARSER.parseFrom(input); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseFrom(input, extensionRegistry); + } + + public static Builder newBuilder() { return Builder.create(); } + public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto prototype) { + return newBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { return newBuilder(this); } + + @java.lang.Override + protected Builder newBuilderForType( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + Builder builder = new Builder(parent); + return builder; + } + /** + * Protobuf type {@code QueryCompleteRequestProto} + */ + public static final class Builder extends + com.google.protobuf.GeneratedMessage.Builder<Builder> + implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProtoOrBuilder { + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteRequestProto_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteRequestProto_fieldAccessorTable + .ensureFieldAccessorsInitialized( + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.Builder.class); + } + + // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.newBuilder() + private Builder() { + maybeForceBuilderInitialization(); + } + + private Builder( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + super(parent); + maybeForceBuilderInitialization(); + } + private void maybeForceBuilderInitialization() { + if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + } + } + private static Builder create() { + return new Builder(); + } + + public Builder clear() { + super.clear(); + queryId_ = ""; + bitField0_ = (bitField0_ & ~0x00000001); + dagName_ = ""; + bitField0_ = (bitField0_ & ~0x00000002); + deleteDelay_ = 0L; + bitField0_ = (bitField0_ & ~0x00000004); + return this; + } + + public Builder clone() { + return create().mergeFrom(buildPartial()); + } + + public com.google.protobuf.Descriptors.Descriptor + getDescriptorForType() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteRequestProto_descriptor; + } + + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto getDefaultInstanceForType() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.getDefaultInstance(); + } + + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto build() { + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException(result); + } + return result; + } + + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto buildPartial() { + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto(this); + int from_bitField0_ = bitField0_; + int to_bitField0_ = 0; + if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + to_bitField0_ |= 0x00000001; + } + result.queryId_ = queryId_; + if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + to_bitField0_ |= 0x00000002; + } + result.dagName_ = dagName_; + if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + to_bitField0_ |= 0x00000004; + } + result.deleteDelay_ = deleteDelay_; + result.bitField0_ = to_bitField0_; + onBuilt(); + return result; + } + + public Builder mergeFrom(com.google.protobuf.Message other) { + if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto) { + return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto)other); + } else { + super.mergeFrom(other); + return this; + } + } + + public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto other) { + if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.getDefaultInstance()) return this; + if (other.hasQueryId()) { + bitField0_ |= 0x00000001; + queryId_ = other.queryId_; + onChanged(); + } + if (other.hasDagName()) { + bitField0_ |= 0x00000002; + dagName_ = other.dagName_; + onChanged(); + } + if (other.hasDeleteDelay()) { + setDeleteDelay(other.getDeleteDelay()); + } + this.mergeUnknownFields(other.getUnknownFields()); + return this; + } + + public final boolean isInitialized() { + return true; + } + + public Builder mergeFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parsedMessage = null; + try { + parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto) e.getUnfinishedMessage(); + throw e; + } finally { + if (parsedMessage != null) { + mergeFrom(parsedMessage); + } + } + return this; + } + private int bitField0_; + + // optional string query_id = 1; + private java.lang.Object queryId_ = ""; + /** + * <code>optional string query_id = 1;</code> + */ + public boolean hasQueryId() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + /** + * <code>optional string query_id = 1;</code> + */ + public java.lang.String getQueryId() { + java.lang.Object ref = queryId_; + if (!(ref instanceof java.lang.String)) { + java.lang.String s = ((com.google.protobuf.ByteString) ref) + .toStringUtf8(); + queryId_ = s; + return s; + } else { + return (java.lang.String) ref; + } + } + /** + * <code>optional string query_id = 1;</code> + */ + public com.google.protobuf.ByteString + getQueryIdBytes() { + java.lang.Object ref = queryId_; + if (ref instanceof String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8( + (java.lang.String) ref); + queryId_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + /** + * <code>optional string query_id = 1;</code> + */ + public Builder setQueryId( + java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000001; + queryId_ = value; + onChanged(); + return this; + } + /** + * <code>optional string query_id = 1;</code> + */ + public Builder clearQueryId() { + bitField0_ = (bitField0_ & ~0x00000001); + queryId_ = getDefaultInstance().getQueryId(); + onChanged(); + return this; + } + /** + * <code>optional string query_id = 1;</code> + */ + public Builder setQueryIdBytes( + com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000001; + queryId_ = value; + onChanged(); + return this; + } + + // optional string dag_name = 2; + private java.lang.Object dagName_ = ""; + /** + * <code>optional string dag_name = 2;</code> + */ + public boolean hasDagName() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + /** + * <code>optional string dag_name = 2;</code> + */ + public java.lang.String getDagName() { + java.lang.Object ref = dagName_; + if (!(ref instanceof java.lang.String)) { + java.lang.String s = ((com.google.protobuf.ByteString) ref) + .toStringUtf8(); + dagName_ = s; + return s; + } else { + return (java.lang.String) ref; + } + } + /** + * <code>optional string dag_name = 2;</code> + */ + public com.google.protobuf.ByteString + getDagNameBytes() { + java.lang.Object ref = dagName_; + if (ref instanceof String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8( + (java.lang.String) ref); + dagName_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + /** + * <code>optional string dag_name = 2;</code> + */ + public Builder setDagName( + java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000002; + dagName_ = value; + onChanged(); + return this; + } + /** + * <code>optional string dag_name = 2;</code> + */ + public Builder clearDagName() { + bitField0_ = (bitField0_ & ~0x00000002); + dagName_ = getDefaultInstance().getDagName(); + onChanged(); + return this; + } + /** + * <code>optional string dag_name = 2;</code> + */ + public Builder setDagNameBytes( + com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000002; + dagName_ = value; + onChanged(); + return this; + } + + // optional int64 delete_delay = 3 [default = 0]; + private long deleteDelay_ ; + /** + * <code>optional int64 delete_delay = 3 [default = 0];</code> + */ + public boolean hasDeleteDelay() { + return ((bitField0_ & 0x00000004) == 0x00000004); + } + /** + * <code>optional int64 delete_delay = 3 [default = 0];</code> + */ + public long getDeleteDelay() { + return deleteDelay_; + } + /** + * <code>optional int64 delete_delay = 3 [default = 0];</code> + */ + public Builder setDeleteDelay(long value) { + bitField0_ |= 0x00000004; + deleteDelay_ = value; + onChanged(); + return this; + } + /** + * <code>optional int64 delete_delay = 3 [default = 0];</code> + */ + public Builder clearDeleteDelay() { + bitField0_ = (bitField0_ & ~0x00000004); + deleteDelay_ = 0L; + onChanged(); + return this; + } + + // @@protoc_insertion_point(builder_scope:QueryCompleteRequestProto) + } + + static { + defaultInstance = new QueryCompleteRequestProto(true); + defaultInstance.initFields(); + } + + // @@protoc_insertion_point(class_scope:QueryCompleteRequestProto) + } + + public interface QueryCompleteResponseProtoOrBuilder + extends com.google.protobuf.MessageOrBuilder { + } + /** + * Protobuf type {@code QueryCompleteResponseProto} + */ + public static final class QueryCompleteResponseProto extends + com.google.protobuf.GeneratedMessage + implements QueryCompleteResponseProtoOrBuilder { + // Use QueryCompleteResponseProto.newBuilder() to construct. + private QueryCompleteResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + super(builder); + this.unknownFields = builder.getUnknownFields(); + } + private QueryCompleteResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + + private static final QueryCompleteResponseProto defaultInstance; + public static QueryCompleteResponseProto getDefaultInstance() { + return defaultInstance; + } + + public QueryCompleteResponseProto getDefaultInstanceForType() { + return defaultInstance; + } + + private final com.google.protobuf.UnknownFieldSet unknownFields; + @java.lang.Override + public final com.google.protobuf.UnknownFieldSet + getUnknownFields() { + return this.unknownFields; + } + private QueryCompleteResponseProto( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + initFields(); + com.google.protobuf.UnknownFieldSet.Builder unknownFields = + com.google.protobuf.UnknownFieldSet.newBuilder(); + try { + boolean done = false; + while (!done) { + int tag = input.readTag(); + switch (tag) { + case 0: + done = true; + break; + default: { + if (!parseUnknownField(input, unknownFields, + extensionRegistry, tag)) { + done = true; + } + break; + } + } + } + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + throw e.setUnfinishedMessage(this); + } catch (java.io.IOException e) { + throw new com.google.protobuf.InvalidProtocolBufferException( + e.getMessage()).setUnfinishedMessage(this); + } finally { + this.unknownFields = unknownFields.build(); + makeExtensionsImmutable(); + } + } + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteResponseProto_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteResponseProto_fieldAccessorTable + .ensureFieldAccessorsInitialized( + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.Builder.class); + } + + public static com.google.protobuf.Parser<QueryCompleteResponseProto> PARSER = + new com.google.protobuf.AbstractParser<QueryCompleteResponseProto>() { + public QueryCompleteResponseProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new QueryCompleteResponseProto(input, extensionRegistry); + } + }; + + @java.lang.Override + public com.google.protobuf.Parser<QueryCompleteResponseProto> getParserForType() { + return PARSER; + } + + private void initFields() { + } + private byte memoizedIsInitialized = -1; + public final boolean isInitialized() { + byte isInitialized = memoizedIsInitialized; + if (isInitialized != -1) return isInitialized == 1; + + memoizedIsInitialized = 1; + return true; + } + + public void writeTo(com.google.protobuf.CodedOutputStream output) + throws java.io.IOException { + getSerializedSize(); + getUnknownFields().writeTo(output); + } + + private int memoizedSerializedSize = -1; + public int getSerializedSize() { + int size = memoizedSerializedSize; + if (size != -1) return size; + + size = 0; + size += getUnknownFields().getSerializedSize(); + memoizedSerializedSize = size; + return size; + } + + private static final long serialVersionUID = 0L; + @java.lang.Override + protected java.lang.Object writeReplace() + throws java.io.ObjectStreamException { + return super.writeReplace(); + } + + @java.lang.Override + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto)) { + return super.equals(obj); + } + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto) obj; + + boolean result = true; + result = result && + getUnknownFields().equals(other.getUnknownFields()); + return result; + } + + private int memoizedHashCode = 0; + @java.lang.Override + public int hashCode() { + if (memoizedHashCode != 0) { + return memoizedHashCode; + } + int hash = 41; + hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (29 * hash) + getUnknownFields().hashCode(); + memoizedHashCode = hash; + return hash; + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom(byte[] data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom( + byte[] data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom(java.io.InputStream input) + throws java.io.IOException { + return PARSER.parseFrom(input); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseFrom(input, extensionRegistry); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseDelimitedFrom(java.io.InputStream input) + throws java.io.IOException { + return PARSER.parseDelimitedFrom(input); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseDelimitedFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseDelimitedFrom(input, extensionRegistry); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom( + com.google.protobuf.CodedInputStream input) + throws java.io.IOException { + return PARSER.parseFrom(input); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseFrom(input, extensionRegistry); + } + + public static Builder newBuilder() { return Builder.create(); } + public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto prototype) { + return newBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { return newBuilder(this); } + + @java.lang.Override + protected Builder newBuilderForType( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + Builder builder = new Builder(parent); + return builder; + } + /** + * Protobuf type {@code QueryCompleteResponseProto} + */ + public static final class Builder extends + com.google.protobuf.GeneratedMessage.Builder<Builder> + implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProtoOrBuilder { + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteResponseProto_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteResponseProto_fieldAccessorTable + .ensureFieldAccessorsInitialized( + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.Builder.class); + } + + // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.newBuilder() + private Builder() { + maybeForceBuilderInitialization(); + } + + private Builder( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + super(parent); + maybeForceBuilderInitialization(); + } + private void maybeForceBuilderInitialization() { + if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + } + } + private static Builder create() { + return new Builder(); + } + + public Builder clear() { + super.clear(); + return this; + } + + public Builder clone() { + return create().mergeFrom(buildPartial()); + } + + public com.google.protobuf.Descriptors.Descriptor + getDescriptorForType() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteResponseProto_descriptor; + } + + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto getDefaultInstanceForType() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.getDefaultInstance(); + } + + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto build() { + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException(result); + } + return result; + } + + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto buildPartial() { + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto(this); + onBuilt(); + return result; + } + + public Builder mergeFrom(com.google.protobuf.Message other) { + if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto) { + return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto)other); + } else { + super.mergeFrom(other); + return this; + } + } + + public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto other) { + if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.getDefaultInstance()) return this; + this.mergeUnknownFields(other.getUnknownFields()); + return this; + } + + public final boolean isInitialized() { + return true; + } + + public Builder mergeFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parsedMessage = null; + try { + parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto) e.getUnfinishedMessage(); + throw e; + } finally { + if (parsedMessage != null) { + mergeFrom(parsedMessage); + } + } + return this; + } + + // @@protoc_insertion_point(builder_scope:QueryCompleteResponseProto) + } + + static { + defaultInstance = new QueryCompleteResponseProto(true); + defaultInstance.initFields(); + } + + // @@protoc_insertion_point(class_scope:QueryCompleteResponseProto) + } + + public interface TerminateFragmentRequestProtoOrBuilder + extends com.google.protobuf.MessageOrBuilder { + + // optional string query_id = 1; + /** + * <code>optional string query_id = 1;</code> + */ + boolean hasQueryId(); + /** + * <code>optional string query_id = 1;</code> + */ + java.lang.String getQueryId(); + /** + * <code>optional string query_id = 1;</code> + */ + com.google.protobuf.ByteString + getQueryIdBytes(); + + // optional string dag_name = 2; + /** + * <code>optional string dag_name = 2;</code> + */ + boolean hasDagName(); + /** + * <code>optional string dag_name = 2;</code> + */ + java.lang.String getDagName(); + /** + * <code>optional string dag_name = 2;</code> + */ + com.google.protobuf.ByteString + getDagNameBytes(); + + // optional int32 dag_attempt_number = 3; + /** + * <code>optional int32 dag_attempt_number = 3;</code> + */ + boolean hasDagAttemptNumber(); + /** + * <code>optional int32 dag_attempt_number = 3;</code> + */ + int getDagAttemptNumber(); + + // optional string vertex_name = 4; + /** + * <code>optional string vertex_name = 4;</code> + */ + boolean hasVertexName(); + /** + * <code>optional string vertex_name = 4;</code> + */ + java.lang.String getVertexName(); + /** + * <code>optional string vertex_name = 4;</code> + */ + com.google.protobuf.ByteString + getVertexNameBytes(); + + // optional int32 fragment_number = 5; + /** + * <code>optional int32 fragment_number = 5;</code> + */ + boolean hasFragmentNumber(); + /** + * <code>optional int32 fragment_number = 5;</code> + */ + int getFragmentNumber(); + + // optional int32 attempt_number = 6; + /** + * <code>optional int32 attempt_number = 6;</code> + */ + boolean hasAttemptNumber(); + /** + * <code>optional int32 attempt_number = 6;</code> + */ + int getAttemptNumber(); + } + /** + * Protobuf type {@code TerminateFragmentRequestProto} + */ + public static final class TerminateFragmentRequestProto extends + com.google.protobuf.GeneratedMessage + implements TerminateFragmentRequestProtoOrBuilder { + // Use TerminateFragmentRequestProto.newBuilder() to construct. + private TerminateFragmentRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + super(builder); + this.unknownFields = builder.getUnknownFields(); + } + private TerminateFragmentRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + + private static final TerminateFragmentRequestProto defaultInstance; + public static TerminateFragmentRequestProto getDefaultInstance() { + return defaultInstance; + } + + public TerminateFragmentRequestProto getDefaultInstanceForType() { + return defaultInstance; + } + + private final com.google.protobuf.UnknownFieldSet unknownFields; + @java.lang.Override + public final com.google.protobuf.UnknownFieldSet + getUnknownFields() { + return this.unknownFields; + } + private TerminateFragmentRequestProto( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + initFields(); + int mutable_bitField0_ = 0; + com.google.protobuf.UnknownFieldSet.Builder unknownFields = + com.google.protobuf.UnknownFieldSet.newBuilder(); + try { + boolean done = false; + while (!done) { + int tag = input.readTag(); + switch (tag) { + case 0: + done = true; + break; + default: { + if (!parseUnknownField(input, unknownFields, + extensionRegistry, tag)) { + done = true; + } + break; + } + case 10: { + bitField0_ |= 0x00000001; + queryId_ = input.readBytes(); + break; + } + case 18: { + bitField0_ |= 0x00000002; + dagName_ = input.readBytes(); + break; + } + case 24: { + bitField0_ |= 0x00000004; + dagAttemptNumber_ = input.readInt32(); + break; + } + case 34: { + bitField0_ |= 0x00000008; + vertexName_ = input.readBytes(); + break; + } + case 40: { + bitField0_ |= 0x00000010; + fragmentNumber_ = input.readInt32(); + break; + } + case 48: { + bitField0_ |= 0x00000020; + attemptNumber_ = input.readInt32(); + break; + } + } + } + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + throw e.setUnfinishedMessage(this); + } catch (java.io.IOException e) { + throw new com.google.protobuf.InvalidProtocolBufferException( + e.getMessage()).setUnfinishedMessage(this); + } finally { + this.unknownFields = unknownFields.build(); + makeExtensionsImmutable(); + } + } + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentRequestProto_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentRequestProto_fieldAccessorTable + .ensureFieldAccessorsInitialized( + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.Builder.class); + } + + public static com.google.protobuf.Parser<TerminateFragmentRequestProto> PARSER = + new com.google.protobuf.AbstractParser<TerminateFragmentRequestProto>() { + public TerminateFragmentRequestProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new TerminateFragmentRequestProto(input, extensionRegistry); + } + }; + + @java.lang.Override + public com.google.protobuf.Parser<TerminateFragmentRequestProto> getParserForType() { + return PARSER; + } + + private int bitField0_; + // optional string query_id = 1; + public static final int QUERY_ID_FIELD_NUMBER = 1; + private java.lang.Object queryId_; + /** + * <code>optional string query_id = 1;</code> + */ + public boolean hasQueryId() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + /** + * <code>optional string query_id = 1;</code> + */ + public java.lang.String getQueryId() { + java.lang.Object ref = queryId_; + if (ref instanceof java.lang.String) { + return (java.lang.String) ref; + } else { + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + queryId_ = s; + } + return s; + } + } + /** + * <code>optional string query_id = 1;</code> + */ + public com.google.protobuf.ByteString + getQueryIdBytes() { + java.lang.Object ref = queryId_; + if (ref instanceof java.lang.String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8( + (java.lang.String) ref); + queryId_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + + // optional string dag_name = 2; + public static final int DAG_NAME_FIELD_NUMBER = 2; + private java.lang.Object dagName_; + /** + * <code>optional string dag_name = 2;</code> + */ + public boolean hasDagName() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + /** + * <code>optional string dag_name = 2;</code> + */ + public java.lang.String getDagName() { + java.lang.Object ref = dagName_; + if (ref instanceof java.lang.String) { + return (java.lang.String) ref; + } else { + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + dagName_ = s; + } + return s; + } + } + /** + * <code>optional string dag_name = 2;</code> + */ + public com.google.protobuf.ByteString + getDagNameBytes() { + java.lang.Object ref = dagName_; + if (ref instanceof java.lang.String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8( + (java.lang.String) ref); + dagName_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + + // optional int32 dag_attempt_number = 3; + public static final int DAG_ATTEMPT_NUMBER_FIELD_NUMBER = 3; + private int dagAttemptNumber_; + /** + * <code>optional int32 dag_attempt_number = 3;</code> + */ + public boolean hasDagAttemptNumber() { + return ((bitField0_ & 0x00000004) == 0x00000004); + } + /** + * <code>optional int32 dag_attempt_number = 3;</code> + */ + public int getDagAttemptNumber() { + return dagAttemptNumber_; + } + + // optional string vertex_name = 4; + public static final int VERTEX_NAME_FIELD_NUMBER = 4; + private java.lang.Object vertexName_; + /** + * <code>optional string vertex_name = 4;</code> + */ + public boolean hasVertexName() { + return ((bitField0_ & 0x00000008) == 0x00000008); + } + /** + * <code>optional string vertex_name = 4;</code> + */ + public java.lang.String getVertexName() { + java.lang.Object ref = vertexName_; + if (ref instanceof java.lang.String) { + return (java.lang.String) ref; + } else { + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + vertexName_ = s; + } + return s; + } + } + /** + * <code>optional string vertex_name = 4;</code> + */ + public com.google.protobuf.ByteString + getVertexNameBytes() { + java.lang.Object ref = vertexName_; + if (ref instanceof java.lang.String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8( + (java.lang.String) ref); + vertexName_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + + // optional int32 fragment_number = 5; + public static final int FRAGMENT_NUMBER_FIELD_NUMBER = 5; + private int fragmentNumber_; + /** + * <code>optional int32 fragment_number = 5;</code> + */ + public boolean hasFragmentNumber() { + return ((bitField0_ & 0x00000010) == 0x00000010); + } + /** + * <code>optional int32 fragment_number = 5;</code> + */ + public int getFragmentNumber() { + return fragmentNumber_; + } + + // optional int32 attempt_number = 6; + public static final int ATTEMPT_NUMBER_FIELD_NUMBER = 6; + private int attemptNumber_; + /** + * <code>optional int32 attempt_number = 6;</code> + */ + public boolean hasAttemptNumber() { + return ((bitField0_ & 0x00000020) == 0x00000020); + } + /** + * <code>optional int32 attempt_number = 6;</code> + */ + public int getAttemptNumber() { + return attemptNumber_; + } + + private void initFields() { + queryId_ = ""; + dagName_ = ""; + dagAttemptNumber_ = 0; + vertexName_ = ""; + fragmentNumber_ = 0; + attemptNumber_ = 0; + } + private byte memoizedIsInitialized = -1; + public final boolean isInitialized() { + byte isInitialized = memoizedIsInitialized; + if (isInitialized != -1) return isInitialized == 1; + + memoizedIsInitialized = 1; + return true; + } + + public void writeTo(com.google.protobuf.CodedOutputStream output) + throws java.io.IOException { + getSerializedSize(); + if (((bitField0_ & 0x00000001) == 0x00000001)) { + output.writeBytes(1, getQueryIdBytes()); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + output.writeBytes(2, getDagNameBytes()); + } + if (((bitField0_ & 0x00000004) == 0x00000004)) { + output.writeInt32(3, dagAttemptNumber_); + } + if (((bitField0_ & 0x00000008) == 0x00000008)) { + output.writeBytes(4, getVertexNameBytes()); + } + if (((bitField0_ & 0x00000010) == 0x00000010)) { + output.writeInt32(5, fragmentNumber_); + } + if (((bitField0_ & 0x00000020) == 0x00000020)) { + output.writeInt32(6, attemptNumber_); + } + getUnknownFields().writeTo(output); + } + + private int memoizedSerializedSize = -1; + public int getSerializedSize() { + int size = memoizedSerializedSize; + if (size != -1) return size; + + size = 0; + if (((bitField0_ & 0x00000001) == 0x00000001)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(1, getQueryIdBytes()); + } + if (((bitField0_ & 0x00000002) == 0x00000002)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(2, getDagNameBytes()); + } + if (((bitField0_ & 0x00000004) == 0x00000004)) { + size += com.google.protobuf.CodedOutputStream + .computeInt32Size(3, dagAttemptNumber_); + } + if (((bitField0_ & 0x00000008) == 0x00000008)) { + size += com.google.protobuf.CodedOutputStream + .computeBytesSize(4, getVertexNameBytes()); + } + if (((bitField0_ & 0x00000010) == 0x00000010)) { + size += com.google.protobuf.CodedOutputStream + .computeInt32Size(5, fragmentNumber_); + } + if (((bitField0_ & 0x00000020) == 0x00000020)) { + size += com.google.protobuf.CodedOutputStream + .computeInt32Size(6, attemptNumber_); + } + size += getUnknownFields().getSerializedSize(); + memoizedSerializedSize = size; + return size; + } + + private static final long serialVersionUID = 0L; + @java.lang.Override + protected java.lang.Object writeReplace() + throws java.io.ObjectStreamException { + return super.writeReplace(); + } + + @java.lang.Override + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto)) { + return super.equals(obj); + } + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto) obj; + + boolean result = true; + result = result && (hasQueryId() == other.hasQueryId()); + if (hasQueryId()) { + result = result && getQueryId() + .equals(other.getQueryId()); + } + result = result && (hasDagName() == other.hasDagName()); + if (hasDagName()) { + result = result && getDagName() + .equals(other.getDagName()); + } + result = result && (hasDagAttemptNumber() == other.hasDagAttemptNumber()); + if (hasDagAttemptNumber()) { + result = result && (getDagAttemptNumber() + == other.getDagAttemptNumber()); + } + result = result && (hasVertexName() == other.hasVertexName()); + if (hasVertexName()) { + result = result && getVertexName() + .equals(other.getVertexName()); + } + result = result && (hasFragmentNumber() == other.hasFragmentNumber()); + if (hasFragmentNumber()) { + result = result && (getFragmentNumber() + == other.getFragmentNumber()); + } + result = result && (hasAttemptNumber() == other.hasAttemptNumber()); + if (hasAttemptNumber()) { + result = result && (getAttemptNumber() + == other.getAttemptNumber()); + } + result = result && + getUnknownFields().equals(other.getUnknownFields()); + return result; + } + + private int memoizedHashCode = 0; + @java.lang.Override + public int hashCode() { + if (memoizedHashCode != 0) { + return memoizedHashCode; + } + int hash = 41; + hash = (19 * hash) + getDescriptorForType().hashCode(); + if (hasQueryId()) { + hash = (37 * hash) + QUERY_ID_FIELD_NUMBER; + hash = (53 * hash) + getQueryId().hashCode(); + } + if (hasDagName()) { + hash = (37 * hash) + DAG_NAME_FIELD_NUMBER; + hash = (53 * hash) + getDagName().hashCode(); + } + if (hasDagAttemptNumber()) { + hash = (37 * hash) + DAG_ATTEMPT_NUMBER_FIELD_NUMBER; + hash = (53 * hash) + getDagAttemptNumber(); + } + if (hasVertexName()) { + hash = (37 * hash) + VERTEX_NAME_FIELD_NUMBER; + hash = (53 * hash) + getVertexName().hashCode(); + } + if (hasFragmentNumber()) { + hash = (37 * hash) + FRAGMENT_NUMBER_FIELD_NUMBER; + hash = (53 * hash) + getFragmentNumber(); + } + if (hasAttemptNumber()) { + hash = (37 * hash) + ATTEMPT_NUMBER_FIELD_NUMBER; + hash = (53 * hash) + getAttemptNumber(); + } + hash = (29 * hash) + getUnknownFields().hashCode(); + memoizedHashCode = hash; + return hash; + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom(byte[] data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom( + byte[] data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom(java.io.InputStream input) + throws java.io.IOException { + return PARSER.parseFrom(input); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseFrom(input, extensionRegistry); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseDelimitedFrom(java.io.InputStream input) + throws java.io.IOException { + return PARSER.parseDelimitedFrom(input); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseDelimitedFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseDelimitedFrom(input, extensionRegistry); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom( + com.google.protobuf.CodedInputStream input) + throws java.io.IOException { + return PARSER.parseFrom(input); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return PARSER.parseFrom(input, extensionRegistry); + } + + public static Builder newBuilder() { return Builder.create(); } + public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto prototype) { + return newBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { return newBuilder(this); } + + @java.lang.Override + protected Builder newBuilderForType( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + Builder builder = new Builder(parent); + return builder; + } + /** + * Protobuf type {@code TerminateFragmentRequestProto} + */ + public static final class Builder extends + com.google.protobuf.GeneratedMessage.Builder<Builder> + implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProtoOrBuilder { + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentRequestProto_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentRequestProto_fieldAccessorTable + .ensureFieldAccessorsInitialized( + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.Builder.class); + } + + // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.newBuilder() + private Builder() { + maybeForceBuilderInitialization(); + } + + private Builder( + com.google.protobuf.GeneratedMessage.BuilderParent parent) { + super(parent); + maybeForceBuilderInitialization(); + } + private void maybeForceBuilderInitialization() { + if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + } + } + private static Builder create() { + return new Builder(); + } + + public Builder clear() { + super.clear(); + queryId_ = ""; + bitField0_ = (bitField0_ & ~0x00000001); + dagName_ = ""; + bitField0_ = (bitField0_ & ~0x00000002); + dagAttemptNumber_ = 0; + bitField0_ = (bitField0_ & ~0x00000004); + vertexName_ = ""; + bitField0_ = (bitField0_ & ~0x00000008); + fragmentNumber_ = 0; + bitField0_ = (bitField0_ & ~0x00000010); + attemptNumber_ = 0; + bitField0_ = (bitField0_ & ~0x00000020); + return this; + } + + public Builder clone() { + return create().mergeFrom(buildPartial()); + } + + public com.google.protobuf.Descriptors.Descriptor + getDescriptorForType() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentRequestProto_descriptor; + } + + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto getDefaultInstanceForType() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.getDefaultInstance(); + } + + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto build() { + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto result = buildPartial(); + if (!result.isInitialized()) { + throw newUninitializedMessageException(result); + } + return result; + } + + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto buildPartial() { + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto(this); + int from_bitField0_ = bitField0_; + int to_bitField0_ = 0; + if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + to_bitField0_ |= 0x00000001; + } + result.queryId_ = queryId_; + if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + to_bitField0_ |= 0x00000002; + } + result.dagName_ = dagName_; + if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + to_bitField0_ |= 0x00000004; + } + result.dagAttemptNumber_ = dagAttemptNumber_; + if (((from_bitField0_ & 0x00000008) == 0x00000008)) { + to_bitField0_ |= 0x00000008; + } + result.vertexName_ = vertexName_; + if (((from_bitField0_ & 0x00000010) == 0x00000010)) { + to_bitField0_ |= 0x00000010; + } + result.fragmentNumber_ = fragmentNumber_; + if (((from_bitField0_ & 0x00000020) == 0x00000020)) { + to_bitField0_ |= 0x00000020; + } + result.attemptNumber_ = attemptNumber_; + result.bitField0_ = to_bitField0_; + onBuilt(); + return result; + } + + public Builder mergeFrom(com.google.protobuf.Message other) { + if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto) { + return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto)other); + } else { + super.mergeFrom(other); + return this; + } + } + + public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto other) { + if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.getDefaultInstance()) return this; + if (other.hasQueryId()) { + bitField0_ |= 0x00000001; + queryId_ = other.queryId_; + onChanged(); + } + if (other.hasDagName()) { + bitField0_ |= 0x00000002; + dagName_ = other.dagName_; + onChanged(); + } + if (other.hasDagAttemptNumber()) { + setDagAttemptNumber(other.getDagAttemptNumber()); + } + if (other.hasVertexName()) { + bitField0_ |= 0x00000008; + vertexName_ = other.vertexName_; + onChanged(); + } + if (other.hasFragmentNumber()) { + setFragmentNumber(other.getFragmentNumber()); + } + if (other.hasAttemptNumber()) { + setAttemptNumber(other.getAttemptNumber()); + } + this.mergeUnknownFields(other.getUnknownFields()); + return this; + } + + public final boolean isInitialized() { + return true; + } + + public Builder mergeFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parsedMessage = null; + try { + parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto) e.getUnfinishedMessage(); + throw e; + } finally { + if (parsedMessage != null) { + mergeFrom(parsedMessage); + } + } + return this; + } + private int bitField0_; + + // optional string query_id = 1; + private java.lang.Object queryId_ = ""; + /** + * <code>optional string query_id = 1;</code> + */ + public boolean hasQueryId() { + return ((bitField0_ & 0x00000001) == 0x00000001); + } + /** + * <code>optional string query_id = 1;</code> + */ + public java.lang.String getQueryId() { + java.lang.Object ref = queryId_; + if (!(ref instanceof java.lang.String)) { + java.lang.String s = ((com.google.protobuf.ByteString) ref) + .toStringUtf8(); + queryId_ = s; + return s; + } else { + return (java.lang.String) ref; + } + } + /** + * <code>optional string query_id = 1;</code> + */ + public com.google.protobuf.ByteString + getQueryIdBytes() { + java.lang.Object ref = queryId_; + if (ref instanceof String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8( + (java.lang.String) ref); + queryId_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + /** + * <code>optional string query_id = 1;</code> + */ + public Builder setQueryId( + java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000001; + queryId_ = value; + onChanged(); + return this; + } + /** + * <code>optional string query_id = 1;</code> + */ + public Builder clearQueryId() { + bitField0_ = (bitField0_ & ~0x00000001); + queryId_ = getDefaultInstance().getQueryId(); + onChanged(); + return this; + } + /** + * <code>optional string query_id = 1;</code> + */ + public Builder setQueryIdBytes( + com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000001; + queryId_ = value; + onChanged(); + return this; + } + + // optional string dag_name = 2; + private java.lang.Object dagName_ = ""; + /** + * <code>optional string dag_name = 2;</code> + */ + public boolean hasDagName() { + return ((bitField0_ & 0x00000002) == 0x00000002); + } + /** + * <code>optional string dag_name = 2;</code> + */ + public java.lang.String getDagName() { + java.lang.Object ref = dagName_; + if (!(ref instanceof java.lang.String)) { + java.lang.String s = ((com.google.protobuf.ByteString) ref) + .toStringUtf8(); + dagName_ = s; + return s; + } else { + return (java.lang.String) ref; + } + } + /** + * <code>optional string dag_name = 2;</code> + */ + public com.google.protobuf.ByteString + getDagNameBytes() { + java.lang.Object ref = dagName_; + if (ref instanceof String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8( + (java.lang.String) ref); + dagName_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + /** + * <code>optional string dag_name = 2;</code> + */ + public Builder setDagName( + java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000002; + dagName_ = value; + onChanged(); + return this; + } + /** + * <code>optional string dag_name = 2;</code> + */ + public Builder clearDagName() { + bitField0_ = (bitField0_ & ~0x00000002); + dagName_ = getDefaultInstance().getDagName(); + onChanged(); + return this; + } + /** + * <code>optional string dag_name = 2;</code> + */ + public Builder setDagNameBytes( + com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000002; + dagName_ = value; + onChanged(); + return this; + } + + // optional int32 dag_attempt_number = 3; + private int dagAttemptNumber_ ; + /** + * <code>optional int32 dag_attempt_number = 3;</code> + */ + public boolean hasDagAttemptNumber() { + return ((bitField0_ & 0x00000004) == 0x00000004); + } + /** + * <code>optional int32 dag_attempt_number = 3;</code> + */ + public int getDagAttemptNumber() { + return dagAttemptNumber_; + } + /** + * <code>optional int32 dag_attempt_number = 3;</code> + */ + public Builder setDagAttemptNumber(int value) { + bitField0_ |= 0x00000004; + dagAttemptNumber_ = value; + onChanged(); + return this; + } + /** + * <code>optional int32 dag_attempt_number = 3;</code> + */ + public Builder clearDagAttemptNumber() { + bitField0_ = (bitField0_ & ~0x00000004); + dagAttemptNumber_ = 0; + onChanged(); + return this; + } + + // optional string vertex_name = 4; + private java.lang.Object vertexName_ = ""; + /** + * <code>optional string vertex_name = 4;</code> + */ + public boolean hasVertexName() { + return ((bitField0_ & 0x00000008) == 0x00000008); + } + /** + * <code>optional string vertex_name = 4;</code> + */ + public java.lang.String getVertexName() { + java.lang.Object ref = vertexName_; + if (!(ref instanceof java.lang.String)) { + java.lang.String s = ((com.google.protobuf.ByteString) ref) + .toStringUtf8(); + vertexName_ = s; + return s; + } else { + return (java.lang.String) ref; + } + } + /** + * <code>optional string vertex_name = 4;</code> + */ + public com.google.protobuf.ByteString + getVertexNameBytes() { + java.lang.Object ref = vertexName_; + if (ref instanceof String) { + com.google.protobuf.ByteString b = + com.google.protobuf.ByteString.copyFromUtf8( + (java.lang.String) ref); + vertexName_ = b; + return b; + } else { + return (com.google.protobuf.ByteString) ref; + } + } + /** + * <code>optional string vertex_name = 4;</code> + */ + public Builder setVertexName( + java.lang.String value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000008; + vertexName_ = value; + onChanged(); + return this; + } + /** + * <code>optional string vertex_name = 4;</code> + */ + public Builder clearVertexName() { + bitField0_ = (bitField0_ & ~0x00000008); + vertexName_ = getDefaultInstance().getVertexName(); + onChanged(); + return this; + } + /** + * <code>optional string vertex_name = 4;</code> + */ + public Builder setVertexNameBytes( + com.google.protobuf.ByteString value) { + if (value == null) { + throw new NullPointerException(); + } + bitField0_ |= 0x00000008; + vertexName_ = value; + onChanged(); + return this; + } + + // optional int32 fragment_number = 5; + private int fragmentNumber_ ; + /** + * <code>optional int32 fragment_number = 5;</code> + */ + public boolean hasFragmentNumber() { + return ((bitField0_ & 0x00000010) == 0x00000010); + } + /** + * <code>optional int32 fragment_number = 5;</code> + */ + public int getFragmentNumber() { + return fragmentNumber_; + } + /** + * <code>optional int32 fragment_number = 5;</code> + */ + public Builder setFragmentNumber(int value) { + bitField0_ |= 0x00000010; + fragmentNumber_ = value; + onChanged(); + return this; + } + /** + * <code>optional int32 fragment_number = 5;</code> + */ + public Builder clearFragmentNumber() { + bitField0_ = (bitField0_ & ~0x00000010); + fragmentNumber_ = 0; + onChanged(); + return this; + } + + // optional int32 attempt_number = 6; + private int attemptNumber_ ; + /** + * <code>optional int32 attempt_number = 6;</code> + */ + public boolean hasAttemptNumber() { + return ((bitField0_ & 0x00000020) == 0x00000020); + } + /** + * <code>optional int32 attempt_number = 6;</code> + */ + public int getAttemptNumber() { + return attemptNumber_; + } + /** + * <code>optional int32 attempt_number = 6;</code> + */ + public Builder setAttemptNumber(int value) { + bitField0_ |= 0x00000020; + attemptNumber_ = value; + onChanged(); + return this; + } + /** + * <code>optional int32 attempt_number = 6;</code> + */ + public Builder clearAttemptNumber() { + bitField0_ = (bitField0_ & ~0x00000020); + attemptNumber_ = 0; + onChanged(); + return this; + } + + // @@protoc_insertion_point(builder_scope:TerminateFragmentRequestProto) + } + + static { + defaultInstance = new TerminateFragmentRequestProto(true); + defaultInstance.initFields(); + } + + // @@protoc_insertion_point(class_scope:TerminateFragmentRequestProto) + } + + public interface TerminateFragmentResponseProtoOrBuilder + extends com.google.protobuf.MessageOrBuilder { + } + /** + * Protobuf type {@code TerminateFragmentResponseProto} + */ + public static final class TerminateFragmentResponseProto extends + com.google.protobuf.GeneratedMessage + implements TerminateFragmentResponseProtoOrBuilder { + // Use TerminateFragmentResponseProto.newBuilder() to construct. + private TerminateFragmentResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + super(builder); + this.unknownFields = builder.getUnknownFields(); + } + private TerminateFragmentResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + + private static final TerminateFragmentResponseProto defaultInstance; + public static TerminateFragmentResponseProto getDefaultInstance() { + return defaultInstance; + } + + public TerminateFragmentResponseProto getDefaultInstanceForType() { + return defaultInstance; + } + + private final com.google.protobuf.UnknownFieldSet unknownFields; + @java.lang.Override + public final com.google.protobuf.UnknownFieldSet + getUnknownFields() { + return this.unknownFields; + } + private TerminateFragmentResponseProto( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + initFields(); + com.google.protobuf.UnknownFieldSet.Builder unknownFields = + com.google.protobuf.UnknownFieldSet.newBuilder(); + try { + boolean done = false; + while (!done) { + int tag = input.readTag(); + switch (tag) { + case 0: + done = true; + break; + default: { + if (!parseUnknownField(input, unknownFields, + extensionRegistry, tag)) { + done = true; + } + break; + } + } + } + } catch (com.google.protobuf.InvalidProtocolBufferException e) { + throw e.setUnfinishedMessage(this); + } catch (java.io.IOException e) { + throw new com.google.protobuf.InvalidProtocolBufferException( + e.getMessage()).setUnfinishedMessage(this); + } finally { + this.unknownFields = unknownFields.build(); + makeExtensionsImmutable(); + } + } + public static final com.google.protobuf.Descriptors.Descriptor + getDescriptor() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentResponseProto_descriptor; + } + + protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + internalGetFieldAccessorTable() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentResponseProto_fieldAccessorTable + .ensureFieldAccessorsInitialized( + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.Builder.class); + } + + public static com.google.protobuf.Parser<TerminateFragmentResponseProto> PARSER = + new com.google.protobuf.AbstractParser<TerminateFragmentResponseProto>() { + public TerminateFragmentResponseProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new TerminateFragmentResponseProto(input, extensionRegistry); + } + }; + + @java.lang.Override + public com.google.protobuf.Parser<TerminateFragmentResponseProto> getParserForType() { + return PARSER; + } + + private void initFields() { + } + private byte memoizedIsInitialized = -1; + public final boolean isInitialized() { + byte isInitialized = memoizedIsInitialized; + if (isInitialized != -1) return isInitialized == 1; + + memoizedIsInitialized = 1; + return true; + } + + public void writeTo(com.google.protobuf.CodedOutputStream output) + throws java.io.IOException { + getSerializedSize(); + getUnknownFields().writeTo(output); + } + + private int memoizedSerializedSize = -1; + public int getSerializedSize() { + int size = memoizedSerializedSize; + if (size != -1) return size; + + size = 0; + size += getUnknownFields().getSerializedSize(); + memoizedSerializedSize = size; + return size; + } + + private static final long serialVersionUID = 0L; + @java.lang.Override + protected java.lang.Object writeReplace() + throws java.io.ObjectStreamException { + return super.writeReplace(); + } + + @java.lang.Override + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto)) { + return super.equals(obj); + } + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto) obj; + + boolean result = true; + result = result && + getUnknownFields().equals(other.getUnknownFields()); + return result; + } + + private int memoizedHashCode = 0; + @java.lang.Override + public int hashCode() { + if (memoizedHashCode != 0) { + return memoizedHashCode; + } + int hash = 41; + hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (29 * hash) + getUnknownFields().hashCode(); + memoizedHashCode = hash; + return hash; + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseFrom(byte[]
<TRUNCATED>