// Generated by the protocol buffer compiler. DO NOT EDIT! // source: llap-common/src/protobuf/LlapDaemonProtocol.proto package org.apache.hadoop.hive.llap.daemon.rpc; public final class LlapDaemonProtocolProtos { private LlapDaemonProtocolProtos() {} public static void registerAllExtensions( com.google.protobuf.ExtensionRegistry registry) { } /** * Protobuf enum {@code SourceStateProto} */ public enum SourceStateProto implements com.google.protobuf.ProtocolMessageEnum { /** * <code>S_SUCCEEDED = 1;</code> */ S_SUCCEEDED(0, 1), /** * <code>S_RUNNING = 2;</code> */ S_RUNNING(1, 2), ; /** * <code>S_SUCCEEDED = 1;</code> */ public static final int S_SUCCEEDED_VALUE = 1; /** * <code>S_RUNNING = 2;</code> */ public static final int S_RUNNING_VALUE = 2; public final int getNumber() { return value; } public static SourceStateProto valueOf(int value) { switch (value) { case 1: return S_SUCCEEDED; case 2: return S_RUNNING; default: return null; } } public static com.google.protobuf.Internal.EnumLiteMap<SourceStateProto> internalGetValueMap() { return internalValueMap; } private static com.google.protobuf.Internal.EnumLiteMap<SourceStateProto> internalValueMap = new com.google.protobuf.Internal.EnumLiteMap<SourceStateProto>() { public SourceStateProto findValueByNumber(int number) { return SourceStateProto.valueOf(number); } }; public final com.google.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { return getDescriptor().getValues().get(index); } public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { return getDescriptor(); } public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.getDescriptor().getEnumTypes().get(0); } private static final SourceStateProto[] VALUES = values(); public static SourceStateProto valueOf( com.google.protobuf.Descriptors.EnumValueDescriptor desc) { if (desc.getType() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "EnumValueDescriptor is not for this type."); } return VALUES[desc.getIndex()]; } private final int index; private final int value; private SourceStateProto(int index, int value) { this.index = index; this.value = value; } // @@protoc_insertion_point(enum_scope:SourceStateProto) } /** * Protobuf enum {@code SubmissionStateProto} */ public enum SubmissionStateProto implements com.google.protobuf.ProtocolMessageEnum { /** * <code>ACCEPTED = 1;</code> */ ACCEPTED(0, 1), /** * <code>REJECTED = 2;</code> */ REJECTED(1, 2), /** * <code>EVICTED_OTHER = 3;</code> */ EVICTED_OTHER(2, 3), ; /** * <code>ACCEPTED = 1;</code> */ public static final int ACCEPTED_VALUE = 1; /** * <code>REJECTED = 2;</code> */ public static final int REJECTED_VALUE = 2; /** * <code>EVICTED_OTHER = 3;</code> */ public static final int EVICTED_OTHER_VALUE = 3; public final int getNumber() { return value; } public static SubmissionStateProto valueOf(int value) { switch (value) { case 1: return ACCEPTED; case 2: return REJECTED; case 3: return EVICTED_OTHER; default: return null; } } public static com.google.protobuf.Internal.EnumLiteMap<SubmissionStateProto> internalGetValueMap() { return internalValueMap; } private static com.google.protobuf.Internal.EnumLiteMap<SubmissionStateProto> internalValueMap = new com.google.protobuf.Internal.EnumLiteMap<SubmissionStateProto>() { public SubmissionStateProto findValueByNumber(int number) { return SubmissionStateProto.valueOf(number); } }; public final com.google.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { return getDescriptor().getValues().get(index); } public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { return getDescriptor(); } public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.getDescriptor().getEnumTypes().get(1); } private static final SubmissionStateProto[] VALUES = values(); public static SubmissionStateProto valueOf( com.google.protobuf.Descriptors.EnumValueDescriptor desc) { if (desc.getType() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "EnumValueDescriptor is not for this type."); } return VALUES[desc.getIndex()]; } private final int index; private final int value; private SubmissionStateProto(int index, int value) { this.index = index; this.value = value; } // @@protoc_insertion_point(enum_scope:SubmissionStateProto) } public interface UserPayloadProtoOrBuilder extends com.google.protobuf.MessageOrBuilder { // optional bytes user_payload = 1; /** * <code>optional bytes user_payload = 1;</code> */ boolean hasUserPayload(); /** * <code>optional bytes user_payload = 1;</code> */ com.google.protobuf.ByteString getUserPayload(); // optional int32 version = 2; /** * <code>optional int32 version = 2;</code> */ boolean hasVersion(); /** * <code>optional int32 version = 2;</code> */ int getVersion(); } /** * Protobuf type {@code UserPayloadProto} */ public static final class UserPayloadProto extends com.google.protobuf.GeneratedMessage implements UserPayloadProtoOrBuilder { // Use UserPayloadProto.newBuilder() to construct. private UserPayloadProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private UserPayloadProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final UserPayloadProto defaultInstance; public static UserPayloadProto getDefaultInstance() { return defaultInstance; } public UserPayloadProto getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private UserPayloadProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { bitField0_ |= 0x00000001; userPayload_ = input.readBytes(); break; } case 16: { bitField0_ |= 0x00000002; version_ = input.readInt32(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder.class); } public static com.google.protobuf.Parser<UserPayloadProto> PARSER = new com.google.protobuf.AbstractParser<UserPayloadProto>() { public UserPayloadProto parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new UserPayloadProto(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<UserPayloadProto> getParserForType() { return PARSER; } private int bitField0_; // optional bytes user_payload = 1; public static final int USER_PAYLOAD_FIELD_NUMBER = 1; private com.google.protobuf.ByteString userPayload_; /** * <code>optional bytes user_payload = 1;</code> */ public boolean hasUserPayload() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional bytes user_payload = 1;</code> */ public com.google.protobuf.ByteString getUserPayload() { return userPayload_; } // optional int32 version = 2; public static final int VERSION_FIELD_NUMBER = 2; private int version_; /** * <code>optional int32 version = 2;</code> */ public boolean hasVersion() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional int32 version = 2;</code> */ public int getVersion() { return version_; } private void initFields() { userPayload_ = com.google.protobuf.ByteString.EMPTY; version_ = 0; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeBytes(1, userPayload_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeInt32(2, version_); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, userPayload_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(2, version_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto) obj; boolean result = true; result = result && (hasUserPayload() == other.hasUserPayload()); if (hasUserPayload()) { result = result && getUserPayload() .equals(other.getUserPayload()); } result = result && (hasVersion() == other.hasVersion()); if (hasVersion()) { result = result && (getVersion() == other.getVersion()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasUserPayload()) { hash = (37 * hash) + USER_PAYLOAD_FIELD_NUMBER; hash = (53 * hash) + getUserPayload().hashCode(); } if (hasVersion()) { hash = (37 * hash) + VERSION_FIELD_NUMBER; hash = (53 * hash) + getVersion(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code UserPayloadProto} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); userPayload_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); version_ = 0; bitField0_ = (bitField0_ & ~0x00000002); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.userPayload_ = userPayload_; if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } result.version_ = version_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance()) return this; if (other.hasUserPayload()) { setUserPayload(other.getUserPayload()); } if (other.hasVersion()) { setVersion(other.getVersion()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // optional bytes user_payload = 1; private com.google.protobuf.ByteString userPayload_ = com.google.protobuf.ByteString.EMPTY; /** * <code>optional bytes user_payload = 1;</code> */ public boolean hasUserPayload() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional bytes user_payload = 1;</code> */ public com.google.protobuf.ByteString getUserPayload() { return userPayload_; } /** * <code>optional bytes user_payload = 1;</code> */ public Builder setUserPayload(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; userPayload_ = value; onChanged(); return this; } /** * <code>optional bytes user_payload = 1;</code> */ public Builder clearUserPayload() { bitField0_ = (bitField0_ & ~0x00000001); userPayload_ = getDefaultInstance().getUserPayload(); onChanged(); return this; } // optional int32 version = 2; private int version_ ; /** * <code>optional int32 version = 2;</code> */ public boolean hasVersion() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional int32 version = 2;</code> */ public int getVersion() { return version_; } /** * <code>optional int32 version = 2;</code> */ public Builder setVersion(int value) { bitField0_ |= 0x00000002; version_ = value; onChanged(); return this; } /** * <code>optional int32 version = 2;</code> */ public Builder clearVersion() { bitField0_ = (bitField0_ & ~0x00000002); version_ = 0; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:UserPayloadProto) } static { defaultInstance = new UserPayloadProto(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:UserPayloadProto) } public interface EntityDescriptorProtoOrBuilder extends com.google.protobuf.MessageOrBuilder { // optional string class_name = 1; /** * <code>optional string class_name = 1;</code> */ boolean hasClassName(); /** * <code>optional string class_name = 1;</code> */ java.lang.String getClassName(); /** * <code>optional string class_name = 1;</code> */ com.google.protobuf.ByteString getClassNameBytes(); // optional .UserPayloadProto user_payload = 2; /** * <code>optional .UserPayloadProto user_payload = 2;</code> */ boolean hasUserPayload(); /** * <code>optional .UserPayloadProto user_payload = 2;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getUserPayload(); /** * <code>optional .UserPayloadProto user_payload = 2;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder getUserPayloadOrBuilder(); // optional bytes history_text = 3; /** * <code>optional bytes history_text = 3;</code> */ boolean hasHistoryText(); /** * <code>optional bytes history_text = 3;</code> */ com.google.protobuf.ByteString getHistoryText(); } /** * Protobuf type {@code EntityDescriptorProto} */ public static final class EntityDescriptorProto extends com.google.protobuf.GeneratedMessage implements EntityDescriptorProtoOrBuilder { // Use EntityDescriptorProto.newBuilder() to construct. private EntityDescriptorProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private EntityDescriptorProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final EntityDescriptorProto defaultInstance; public static EntityDescriptorProto getDefaultInstance() { return defaultInstance; } public EntityDescriptorProto getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private EntityDescriptorProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { bitField0_ |= 0x00000001; className_ = input.readBytes(); break; } case 18: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder subBuilder = null; if (((bitField0_ & 0x00000002) == 0x00000002)) { subBuilder = userPayload_.toBuilder(); } userPayload_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(userPayload_); userPayload_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000002; break; } case 26: { bitField0_ |= 0x00000004; historyText_ = input.readBytes(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder.class); } public static com.google.protobuf.Parser<EntityDescriptorProto> PARSER = new com.google.protobuf.AbstractParser<EntityDescriptorProto>() { public EntityDescriptorProto parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new EntityDescriptorProto(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<EntityDescriptorProto> getParserForType() { return PARSER; } private int bitField0_; // optional string class_name = 1; public static final int CLASS_NAME_FIELD_NUMBER = 1; private java.lang.Object className_; /** * <code>optional string class_name = 1;</code> */ public boolean hasClassName() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional string class_name = 1;</code> */ public java.lang.String getClassName() { java.lang.Object ref = className_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { className_ = s; } return s; } } /** * <code>optional string class_name = 1;</code> */ public com.google.protobuf.ByteString getClassNameBytes() { java.lang.Object ref = className_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); className_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } // optional .UserPayloadProto user_payload = 2; public static final int USER_PAYLOAD_FIELD_NUMBER = 2; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto userPayload_; /** * <code>optional .UserPayloadProto user_payload = 2;</code> */ public boolean hasUserPayload() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional .UserPayloadProto user_payload = 2;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getUserPayload() { return userPayload_; } /** * <code>optional .UserPayloadProto user_payload = 2;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder getUserPayloadOrBuilder() { return userPayload_; } // optional bytes history_text = 3; public static final int HISTORY_TEXT_FIELD_NUMBER = 3; private com.google.protobuf.ByteString historyText_; /** * <code>optional bytes history_text = 3;</code> */ public boolean hasHistoryText() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional bytes history_text = 3;</code> */ public com.google.protobuf.ByteString getHistoryText() { return historyText_; } private void initFields() { className_ = ""; userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance(); historyText_ = com.google.protobuf.ByteString.EMPTY; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeBytes(1, getClassNameBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeMessage(2, userPayload_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { output.writeBytes(3, historyText_); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, getClassNameBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(2, userPayload_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(3, historyText_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto) obj; boolean result = true; result = result && (hasClassName() == other.hasClassName()); if (hasClassName()) { result = result && getClassName() .equals(other.getClassName()); } result = result && (hasUserPayload() == other.hasUserPayload()); if (hasUserPayload()) { result = result && getUserPayload() .equals(other.getUserPayload()); } result = result && (hasHistoryText() == other.hasHistoryText()); if (hasHistoryText()) { result = result && getHistoryText() .equals(other.getHistoryText()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasClassName()) { hash = (37 * hash) + CLASS_NAME_FIELD_NUMBER; hash = (53 * hash) + getClassName().hashCode(); } if (hasUserPayload()) { hash = (37 * hash) + USER_PAYLOAD_FIELD_NUMBER; hash = (53 * hash) + getUserPayload().hashCode(); } if (hasHistoryText()) { hash = (37 * hash) + HISTORY_TEXT_FIELD_NUMBER; hash = (53 * hash) + getHistoryText().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code EntityDescriptorProto} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getUserPayloadFieldBuilder(); } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); className_ = ""; bitField0_ = (bitField0_ & ~0x00000001); if (userPayloadBuilder_ == null) { userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance(); } else { userPayloadBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); historyText_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000004); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.className_ = className_; if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } if (userPayloadBuilder_ == null) { result.userPayload_ = userPayload_; } else { result.userPayload_ = userPayloadBuilder_.build(); } if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000004; } result.historyText_ = historyText_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) return this; if (other.hasClassName()) { bitField0_ |= 0x00000001; className_ = other.className_; onChanged(); } if (other.hasUserPayload()) { mergeUserPayload(other.getUserPayload()); } if (other.hasHistoryText()) { setHistoryText(other.getHistoryText()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // optional string class_name = 1; private java.lang.Object className_ = ""; /** * <code>optional string class_name = 1;</code> */ public boolean hasClassName() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional string class_name = 1;</code> */ public java.lang.String getClassName() { java.lang.Object ref = className_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); className_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>optional string class_name = 1;</code> */ public com.google.protobuf.ByteString getClassNameBytes() { java.lang.Object ref = className_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); className_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>optional string class_name = 1;</code> */ public Builder setClassName( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; className_ = value; onChanged(); return this; } /** * <code>optional string class_name = 1;</code> */ public Builder clearClassName() { bitField0_ = (bitField0_ & ~0x00000001); className_ = getDefaultInstance().getClassName(); onChanged(); return this; } /** * <code>optional string class_name = 1;</code> */ public Builder setClassNameBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; className_ = value; onChanged(); return this; } // optional .UserPayloadProto user_payload = 2; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder> userPayloadBuilder_; /** * <code>optional .UserPayloadProto user_payload = 2;</code> */ public boolean hasUserPayload() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional .UserPayloadProto user_payload = 2;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getUserPayload() { if (userPayloadBuilder_ == null) { return userPayload_; } else { return userPayloadBuilder_.getMessage(); } } /** * <code>optional .UserPayloadProto user_payload = 2;</code> */ public Builder setUserPayload(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto value) { if (userPayloadBuilder_ == null) { if (value == null) { throw new NullPointerException(); } userPayload_ = value; onChanged(); } else { userPayloadBuilder_.setMessage(value); } bitField0_ |= 0x00000002; return this; } /** * <code>optional .UserPayloadProto user_payload = 2;</code> */ public Builder setUserPayload( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder builderForValue) { if (userPayloadBuilder_ == null) { userPayload_ = builderForValue.build(); onChanged(); } else { userPayloadBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000002; return this; } /** * <code>optional .UserPayloadProto user_payload = 2;</code> */ public Builder mergeUserPayload(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto value) { if (userPayloadBuilder_ == null) { if (((bitField0_ & 0x00000002) == 0x00000002) && userPayload_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance()) { userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.newBuilder(userPayload_).mergeFrom(value).buildPartial(); } else { userPayload_ = value; } onChanged(); } else { userPayloadBuilder_.mergeFrom(value); } bitField0_ |= 0x00000002; return this; } /** * <code>optional .UserPayloadProto user_payload = 2;</code> */ public Builder clearUserPayload() { if (userPayloadBuilder_ == null) { userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance(); onChanged(); } else { userPayloadBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } /** * <code>optional .UserPayloadProto user_payload = 2;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder getUserPayloadBuilder() { bitField0_ |= 0x00000002; onChanged(); return getUserPayloadFieldBuilder().getBuilder(); } /** * <code>optional .UserPayloadProto user_payload = 2;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder getUserPayloadOrBuilder() { if (userPayloadBuilder_ != null) { return userPayloadBuilder_.getMessageOrBuilder(); } else { return userPayload_; } } /** * <code>optional .UserPayloadProto user_payload = 2;</code> */ private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder> getUserPayloadFieldBuilder() { if (userPayloadBuilder_ == null) { userPayloadBuilder_ = new com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder>( userPayload_, getParentForChildren(), isClean()); userPayload_ = null; } return userPayloadBuilder_; } // optional bytes history_text = 3; private com.google.protobuf.ByteString historyText_ = com.google.protobuf.ByteString.EMPTY; /** * <code>optional bytes history_text = 3;</code> */ public boolean hasHistoryText() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional bytes history_text = 3;</code> */ public com.google.protobuf.ByteString getHistoryText() { return historyText_; } /** * <code>optional bytes history_text = 3;</code> */ public Builder setHistoryText(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000004; historyText_ = value; onChanged(); return this; } /** * <code>optional bytes history_text = 3;</code> */ public Builder clearHistoryText() { bitField0_ = (bitField0_ & ~0x00000004); historyText_ = getDefaultInstance().getHistoryText(); onChanged(); return this; } // @@protoc_insertion_point(builder_scope:EntityDescriptorProto) } static { defaultInstance = new EntityDescriptorProto(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:EntityDescriptorProto) } public interface IOSpecProtoOrBuilder extends com.google.protobuf.MessageOrBuilder { // optional string connected_vertex_name = 1; /** * <code>optional string connected_vertex_name = 1;</code> */ boolean hasConnectedVertexName(); /** * <code>optional string connected_vertex_name = 1;</code> */ java.lang.String getConnectedVertexName(); /** * <code>optional string connected_vertex_name = 1;</code> */ com.google.protobuf.ByteString getConnectedVertexNameBytes(); // optional .EntityDescriptorProto io_descriptor = 2; /** * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> */ boolean hasIoDescriptor(); /** * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getIoDescriptor(); /** * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getIoDescriptorOrBuilder(); // optional int32 physical_edge_count = 3; /** * <code>optional int32 physical_edge_count = 3;</code> */ boolean hasPhysicalEdgeCount(); /** * <code>optional int32 physical_edge_count = 3;</code> */ int getPhysicalEdgeCount(); } /** * Protobuf type {@code IOSpecProto} */ public static final class IOSpecProto extends com.google.protobuf.GeneratedMessage implements IOSpecProtoOrBuilder { // Use IOSpecProto.newBuilder() to construct. private IOSpecProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private IOSpecProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final IOSpecProto defaultInstance; public static IOSpecProto getDefaultInstance() { return defaultInstance; } public IOSpecProto getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IOSpecProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { bitField0_ |= 0x00000001; connectedVertexName_ = input.readBytes(); break; } case 18: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder subBuilder = null; if (((bitField0_ & 0x00000002) == 0x00000002)) { subBuilder = ioDescriptor_.toBuilder(); } ioDescriptor_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(ioDescriptor_); ioDescriptor_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000002; break; } case 24: { bitField0_ |= 0x00000004; physicalEdgeCount_ = input.readInt32(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder.class); } public static com.google.protobuf.Parser<IOSpecProto> PARSER = new com.google.protobuf.AbstractParser<IOSpecProto>() { public IOSpecProto parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new IOSpecProto(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<IOSpecProto> getParserForType() { return PARSER; } private int bitField0_; // optional string connected_vertex_name = 1; public static final int CONNECTED_VERTEX_NAME_FIELD_NUMBER = 1; private java.lang.Object connectedVertexName_; /** * <code>optional string connected_vertex_name = 1;</code> */ public boolean hasConnectedVertexName() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional string connected_vertex_name = 1;</code> */ public java.lang.String getConnectedVertexName() { java.lang.Object ref = connectedVertexName_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { connectedVertexName_ = s; } return s; } } /** * <code>optional string connected_vertex_name = 1;</code> */ public com.google.protobuf.ByteString getConnectedVertexNameBytes() { java.lang.Object ref = connectedVertexName_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); connectedVertexName_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } // optional .EntityDescriptorProto io_descriptor = 2; public static final int IO_DESCRIPTOR_FIELD_NUMBER = 2; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto ioDescriptor_; /** * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> */ public boolean hasIoDescriptor() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getIoDescriptor() { return ioDescriptor_; } /** * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getIoDescriptorOrBuilder() { return ioDescriptor_; } // optional int32 physical_edge_count = 3; public static final int PHYSICAL_EDGE_COUNT_FIELD_NUMBER = 3; private int physicalEdgeCount_; /** * <code>optional int32 physical_edge_count = 3;</code> */ public boolean hasPhysicalEdgeCount() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional int32 physical_edge_count = 3;</code> */ public int getPhysicalEdgeCount() { return physicalEdgeCount_; } private void initFields() { connectedVertexName_ = ""; ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); physicalEdgeCount_ = 0; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeBytes(1, getConnectedVertexNameBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeMessage(2, ioDescriptor_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { output.writeInt32(3, physicalEdgeCount_); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, getConnectedVertexNameBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(2, ioDescriptor_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(3, physicalEdgeCount_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto) obj; boolean result = true; result = result && (hasConnectedVertexName() == other.hasConnectedVertexName()); if (hasConnectedVertexName()) { result = result && getConnectedVertexName() .equals(other.getConnectedVertexName()); } result = result && (hasIoDescriptor() == other.hasIoDescriptor()); if (hasIoDescriptor()) { result = result && getIoDescriptor() .equals(other.getIoDescriptor()); } result = result && (hasPhysicalEdgeCount() == other.hasPhysicalEdgeCount()); if (hasPhysicalEdgeCount()) { result = result && (getPhysicalEdgeCount() == other.getPhysicalEdgeCount()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasConnectedVertexName()) { hash = (37 * hash) + CONNECTED_VERTEX_NAME_FIELD_NUMBER; hash = (53 * hash) + getConnectedVertexName().hashCode(); } if (hasIoDescriptor()) { hash = (37 * hash) + IO_DESCRIPTOR_FIELD_NUMBER; hash = (53 * hash) + getIoDescriptor().hashCode(); } if (hasPhysicalEdgeCount()) { hash = (37 * hash) + PHYSICAL_EDGE_COUNT_FIELD_NUMBER; hash = (53 * hash) + getPhysicalEdgeCount(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code IOSpecProto} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getIoDescriptorFieldBuilder(); } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); connectedVertexName_ = ""; bitField0_ = (bitField0_ & ~0x00000001); if (ioDescriptorBuilder_ == null) { ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); } else { ioDescriptorBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); physicalEdgeCount_ = 0; bitField0_ = (bitField0_ & ~0x00000004); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.connectedVertexName_ = connectedVertexName_; if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } if (ioDescriptorBuilder_ == null) { result.ioDescriptor_ = ioDescriptor_; } else { result.ioDescriptor_ = ioDescriptorBuilder_.build(); } if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000004; } result.physicalEdgeCount_ = physicalEdgeCount_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.getDefaultInstance()) return this; if (other.hasConnectedVertexName()) { bitField0_ |= 0x00000001; connectedVertexName_ = other.connectedVertexName_; onChanged(); } if (other.hasIoDescriptor()) { mergeIoDescriptor(other.getIoDescriptor()); } if (other.hasPhysicalEdgeCount()) { setPhysicalEdgeCount(other.getPhysicalEdgeCount()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // optional string connected_vertex_name = 1; private java.lang.Object connectedVertexName_ = ""; /** * <code>optional string connected_vertex_name = 1;</code> */ public boolean hasConnectedVertexName() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional string connected_vertex_name = 1;</code> */ public java.lang.String getConnectedVertexName() { java.lang.Object ref = connectedVertexName_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); connectedVertexName_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>optional string connected_vertex_name = 1;</code> */ public com.google.protobuf.ByteString getConnectedVertexNameBytes() { java.lang.Object ref = connectedVertexName_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); connectedVertexName_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>optional string connected_vertex_name = 1;</code> */ public Builder setConnectedVertexName( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; connectedVertexName_ = value; onChanged(); return this; } /** * <code>optional string connected_vertex_name = 1;</code> */ public Builder clearConnectedVertexName() { bitField0_ = (bitField0_ & ~0x00000001); connectedVertexName_ = getDefaultInstance().getConnectedVertexName(); onChanged(); return this; } /** * <code>optional string connected_vertex_name = 1;</code> */ public Builder setConnectedVertexNameBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; connectedVertexName_ = value; onChanged(); return this; } // optional .EntityDescriptorProto io_descriptor = 2; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> ioDescriptorBuilder_; /** * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> */ public boolean hasIoDescriptor() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getIoDescriptor() { if (ioDescriptorBuilder_ == null) { return ioDescriptor_; } else { return ioDescriptorBuilder_.getMessage(); } } /** * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> */ public Builder setIoDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) { if (ioDescriptorBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ioDescriptor_ = value; onChanged(); } else { ioDescriptorBuilder_.setMessage(value); } bitField0_ |= 0x00000002; return this; } /** * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> */ public Builder setIoDescriptor( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder builderForValue) { if (ioDescriptorBuilder_ == null) { ioDescriptor_ = builderForValue.build(); onChanged(); } else { ioDescriptorBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000002; return this; } /** * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> */ public Builder mergeIoDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) { if (ioDescriptorBuilder_ == null) { if (((bitField0_ & 0x00000002) == 0x00000002) && ioDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) { ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.newBuilder(ioDescriptor_).mergeFrom(value).buildPartial(); } else { ioDescriptor_ = value; } onChanged(); } else { ioDescriptorBuilder_.mergeFrom(value); } bitField0_ |= 0x00000002; return this; } /** * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> */ public Builder clearIoDescriptor() { if (ioDescriptorBuilder_ == null) { ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); onChanged(); } else { ioDescriptorBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } /** * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder getIoDescriptorBuilder() { bitField0_ |= 0x00000002; onChanged(); return getIoDescriptorFieldBuilder().getBuilder(); } /** * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getIoDescriptorOrBuilder() { if (ioDescriptorBuilder_ != null) { return ioDescriptorBuilder_.getMessageOrBuilder(); } else { return ioDescriptor_; } } /** * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> */ private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> getIoDescriptorFieldBuilder() { if (ioDescriptorBuilder_ == null) { ioDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder>( ioDescriptor_, getParentForChildren(), isClean()); ioDescriptor_ = null; } return ioDescriptorBuilder_; } // optional int32 physical_edge_count = 3; private int physicalEdgeCount_ ; /** * <code>optional int32 physical_edge_count = 3;</code> */ public boolean hasPhysicalEdgeCount() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional int32 physical_edge_count = 3;</code> */ public int getPhysicalEdgeCount() { return physicalEdgeCount_; } /** * <code>optional int32 physical_edge_count = 3;</code> */ public Builder setPhysicalEdgeCount(int value) { bitField0_ |= 0x00000004; physicalEdgeCount_ = value; onChanged(); return this; } /** * <code>optional int32 physical_edge_count = 3;</code> */ public Builder clearPhysicalEdgeCount() { bitField0_ = (bitField0_ & ~0x00000004); physicalEdgeCount_ = 0; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:IOSpecProto) } static { defaultInstance = new IOSpecProto(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:IOSpecProto) } public interface GroupInputSpecProtoOrBuilder extends com.google.protobuf.MessageOrBuilder { // optional string group_name = 1; /** * <code>optional string group_name = 1;</code> */ boolean hasGroupName(); /** * <code>optional string group_name = 1;</code> */ java.lang.String getGroupName(); /** * <code>optional string group_name = 1;</code> */ com.google.protobuf.ByteString getGroupNameBytes(); // repeated string group_vertices = 2; /** * <code>repeated string group_vertices = 2;</code> */ java.util.List<java.lang.String> getGroupVerticesList(); /** * <code>repeated string group_vertices = 2;</code> */ int getGroupVerticesCount(); /** * <code>repeated string group_vertices = 2;</code> */ java.lang.String getGroupVertices(int index); /** * <code>repeated string group_vertices = 2;</code> */ com.google.protobuf.ByteString getGroupVerticesBytes(int index); // optional .EntityDescriptorProto merged_input_descriptor = 3; /** * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code> */ boolean hasMergedInputDescriptor(); /** * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getMergedInputDescriptor(); /** * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getMergedInputDescriptorOrBuilder(); } /** * Protobuf type {@code GroupInputSpecProto} */ public static final class GroupInputSpecProto extends com.google.protobuf.GeneratedMessage implements GroupInputSpecProtoOrBuilder { // Use GroupInputSpecProto.newBuilder() to construct. private GroupInputSpecProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private GroupInputSpecProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final GroupInputSpecProto defaultInstance; public static GroupInputSpecProto getDefaultInstance() { return defaultInstance; } public GroupInputSpecProto getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GroupInputSpecProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { bitField0_ |= 0x00000001; groupName_ = input.readBytes(); break; } case 18: { if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) { groupVertices_ = new com.google.protobuf.LazyStringArrayList(); mutable_bitField0_ |= 0x00000002; } groupVertices_.add(input.readBytes()); break; } case 26: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder subBuilder = null; if (((bitField0_ & 0x00000002) == 0x00000002)) { subBuilder = mergedInputDescriptor_.toBuilder(); } mergedInputDescriptor_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(mergedInputDescriptor_); mergedInputDescriptor_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000002; break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) { groupVertices_ = new com.google.protobuf.UnmodifiableLazyStringList(groupVertices_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder.class); } public static com.google.protobuf.Parser<GroupInputSpecProto> PARSER = new com.google.protobuf.AbstractParser<GroupInputSpecProto>() { public GroupInputSpecProto parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new GroupInputSpecProto(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<GroupInputSpecProto> getParserForType() { return PARSER; } private int bitField0_; // optional string group_name = 1; public static final int GROUP_NAME_FIELD_NUMBER = 1; private java.lang.Object groupName_; /** * <code>optional string group_name = 1;</code> */ public boolean hasGroupName() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional string group_name = 1;</code> */ public java.lang.String getGroupName() { java.lang.Object ref = groupName_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { groupName_ = s; } return s; } } /** * <code>optional string group_name = 1;</code> */ public com.google.protobuf.ByteString getGroupNameBytes() { java.lang.Object ref = groupName_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); groupName_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } // repeated string group_vertices = 2; public static final int GROUP_VERTICES_FIELD_NUMBER = 2; private com.google.protobuf.LazyStringList groupVertices_; /** * <code>repeated string group_vertices = 2;</code> */ public java.util.List<java.lang.String> getGroupVerticesList() { return groupVertices_; } /** * <code>repeated string group_vertices = 2;</code> */ public int getGroupVerticesCount() { return groupVertices_.size(); } /** * <code>repeated string group_vertices = 2;</code> */ public java.lang.String getGroupVertices(int index) { return groupVertices_.get(index); } /** * <code>repeated string group_vertices = 2;</code> */ public com.google.protobuf.ByteString getGroupVerticesBytes(int index) { return groupVertices_.getByteString(index); } // optional .EntityDescriptorProto merged_input_descriptor = 3; public static final int MERGED_INPUT_DESCRIPTOR_FIELD_NUMBER = 3; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto mergedInputDescriptor_; /** * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code> */ public boolean hasMergedInputDescriptor() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getMergedInputDescriptor() { return mergedInputDescriptor_; } /** * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getMergedInputDescriptorOrBuilder() { return mergedInputDescriptor_; } private void initFields() { groupName_ = ""; groupVertices_ = com.google.protobuf.LazyStringArrayList.EMPTY; mergedInputDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeBytes(1, getGroupNameBytes()); } for (int i = 0; i < groupVertices_.size(); i++) { output.writeBytes(2, groupVertices_.getByteString(i)); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeMessage(3, mergedInputDescriptor_); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, getGroupNameBytes()); } { int dataSize = 0; for (int i = 0; i < groupVertices_.size(); i++) { dataSize += com.google.protobuf.CodedOutputStream .computeBytesSizeNoTag(groupVertices_.getByteString(i)); } size += dataSize; size += 1 * getGroupVerticesList().size(); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(3, mergedInputDescriptor_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto) obj; boolean result = true; result = result && (hasGroupName() == other.hasGroupName()); if (hasGroupName()) { result = result && getGroupName() .equals(other.getGroupName()); } result = result && getGroupVerticesList() .equals(other.getGroupVerticesList()); result = result && (hasMergedInputDescriptor() == other.hasMergedInputDescriptor()); if (hasMergedInputDescriptor()) { result = result && getMergedInputDescriptor() .equals(other.getMergedInputDescriptor()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasGroupName()) { hash = (37 * hash) + GROUP_NAME_FIELD_NUMBER; hash = (53 * hash) + getGroupName().hashCode(); } if (getGroupVerticesCount() > 0) { hash = (37 * hash) + GROUP_VERTICES_FIELD_NUMBER; hash = (53 * hash) + getGroupVerticesList().hashCode(); } if (hasMergedInputDescriptor()) { hash = (37 * hash) + MERGED_INPUT_DESCRIPTOR_FIELD_NUMBER; hash = (53 * hash) + getMergedInputDescriptor().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code GroupInputSpecProto} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getMergedInputDescriptorFieldBuilder(); } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); groupName_ = ""; bitField0_ = (bitField0_ & ~0x00000001); groupVertices_ = com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); if (mergedInputDescriptorBuilder_ == null) { mergedInputDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); } else { mergedInputDescriptorBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000004); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.groupName_ = groupName_; if (((bitField0_ & 0x00000002) == 0x00000002)) { groupVertices_ = new com.google.protobuf.UnmodifiableLazyStringList( groupVertices_); bitField0_ = (bitField0_ & ~0x00000002); } result.groupVertices_ = groupVertices_; if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000002; } if (mergedInputDescriptorBuilder_ == null) { result.mergedInputDescriptor_ = mergedInputDescriptor_; } else { result.mergedInputDescriptor_ = mergedInputDescriptorBuilder_.build(); } result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.getDefaultInstance()) return this; if (other.hasGroupName()) { bitField0_ |= 0x00000001; groupName_ = other.groupName_; onChanged(); } if (!other.groupVertices_.isEmpty()) { if (groupVertices_.isEmpty()) { groupVertices_ = other.groupVertices_; bitField0_ = (bitField0_ & ~0x00000002); } else { ensureGroupVerticesIsMutable(); groupVertices_.addAll(other.groupVertices_); } onChanged(); } if (other.hasMergedInputDescriptor()) { mergeMergedInputDescriptor(other.getMergedInputDescriptor()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // optional string group_name = 1; private java.lang.Object groupName_ = ""; /** * <code>optional string group_name = 1;</code> */ public boolean hasGroupName() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional string group_name = 1;</code> */ public java.lang.String getGroupName() { java.lang.Object ref = groupName_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); groupName_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>optional string group_name = 1;</code> */ public com.google.protobuf.ByteString getGroupNameBytes() { java.lang.Object ref = groupName_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); groupName_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>optional string group_name = 1;</code> */ public Builder setGroupName( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; groupName_ = value; onChanged(); return this; } /** * <code>optional string group_name = 1;</code> */ public Builder clearGroupName() { bitField0_ = (bitField0_ & ~0x00000001); groupName_ = getDefaultInstance().getGroupName(); onChanged(); return this; } /** * <code>optional string group_name = 1;</code> */ public Builder setGroupNameBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; groupName_ = value; onChanged(); return this; } // repeated string group_vertices = 2; private com.google.protobuf.LazyStringList groupVertices_ = com.google.protobuf.LazyStringArrayList.EMPTY; private void ensureGroupVerticesIsMutable() { if (!((bitField0_ & 0x00000002) == 0x00000002)) { groupVertices_ = new com.google.protobuf.LazyStringArrayList(groupVertices_); bitField0_ |= 0x00000002; } } /** * <code>repeated string group_vertices = 2;</code> */ public java.util.List<java.lang.String> getGroupVerticesList() { return java.util.Collections.unmodifiableList(groupVertices_); } /** * <code>repeated string group_vertices = 2;</code> */ public int getGroupVerticesCount() { return groupVertices_.size(); } /** * <code>repeated string group_vertices = 2;</code> */ public java.lang.String getGroupVertices(int index) { return groupVertices_.get(index); } /** * <code>repeated string group_vertices = 2;</code> */ public com.google.protobuf.ByteString getGroupVerticesBytes(int index) { return groupVertices_.getByteString(index); } /** * <code>repeated string group_vertices = 2;</code> */ public Builder setGroupVertices( int index, java.lang.String value) { if (value == null) { throw new NullPointerException(); } ensureGroupVerticesIsMutable(); groupVertices_.set(index, value); onChanged(); return this; } /** * <code>repeated string group_vertices = 2;</code> */ public Builder addGroupVertices( java.lang.String value) { if (value == null) { throw new NullPointerException(); } ensureGroupVerticesIsMutable(); groupVertices_.add(value); onChanged(); return this; } /** * <code>repeated string group_vertices = 2;</code> */ public Builder addAllGroupVertices( java.lang.Iterable<java.lang.String> values) { ensureGroupVerticesIsMutable(); super.addAll(values, groupVertices_); onChanged(); return this; } /** * <code>repeated string group_vertices = 2;</code> */ public Builder clearGroupVertices() { groupVertices_ = com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); onChanged(); return this; } /** * <code>repeated string group_vertices = 2;</code> */ public Builder addGroupVerticesBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } ensureGroupVerticesIsMutable(); groupVertices_.add(value); onChanged(); return this; } // optional .EntityDescriptorProto merged_input_descriptor = 3; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto mergedInputDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> mergedInputDescriptorBuilder_; /** * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code> */ public boolean hasMergedInputDescriptor() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getMergedInputDescriptor() { if (mergedInputDescriptorBuilder_ == null) { return mergedInputDescriptor_; } else { return mergedInputDescriptorBuilder_.getMessage(); } } /** * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code> */ public Builder setMergedInputDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) { if (mergedInputDescriptorBuilder_ == null) { if (value == null) { throw new NullPointerException(); } mergedInputDescriptor_ = value; onChanged(); } else { mergedInputDescriptorBuilder_.setMessage(value); } bitField0_ |= 0x00000004; return this; } /** * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code> */ public Builder setMergedInputDescriptor( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder builderForValue) { if (mergedInputDescriptorBuilder_ == null) { mergedInputDescriptor_ = builderForValue.build(); onChanged(); } else { mergedInputDescriptorBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000004; return this; } /** * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code> */ public Builder mergeMergedInputDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) { if (mergedInputDescriptorBuilder_ == null) { if (((bitField0_ & 0x00000004) == 0x00000004) && mergedInputDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) { mergedInputDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.newBuilder(mergedInputDescriptor_).mergeFrom(value).buildPartial(); } else { mergedInputDescriptor_ = value; } onChanged(); } else { mergedInputDescriptorBuilder_.mergeFrom(value); } bitField0_ |= 0x00000004; return this; } /** * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code> */ public Builder clearMergedInputDescriptor() { if (mergedInputDescriptorBuilder_ == null) { mergedInputDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); onChanged(); } else { mergedInputDescriptorBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000004); return this; } /** * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder getMergedInputDescriptorBuilder() { bitField0_ |= 0x00000004; onChanged(); return getMergedInputDescriptorFieldBuilder().getBuilder(); } /** * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getMergedInputDescriptorOrBuilder() { if (mergedInputDescriptorBuilder_ != null) { return mergedInputDescriptorBuilder_.getMessageOrBuilder(); } else { return mergedInputDescriptor_; } } /** * <code>optional .EntityDescriptorProto merged_input_descriptor = 3;</code> */ private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> getMergedInputDescriptorFieldBuilder() { if (mergedInputDescriptorBuilder_ == null) { mergedInputDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder>( mergedInputDescriptor_, getParentForChildren(), isClean()); mergedInputDescriptor_ = null; } return mergedInputDescriptorBuilder_; } // @@protoc_insertion_point(builder_scope:GroupInputSpecProto) } static { defaultInstance = new GroupInputSpecProto(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:GroupInputSpecProto) } public interface SignableVertexSpecOrBuilder extends com.google.protobuf.MessageOrBuilder { // optional string user = 1; /** * <code>optional string user = 1;</code> */ boolean hasUser(); /** * <code>optional string user = 1;</code> */ java.lang.String getUser(); /** * <code>optional string user = 1;</code> */ com.google.protobuf.ByteString getUserBytes(); // optional int64 signatureKeyId = 2; /** * <code>optional int64 signatureKeyId = 2;</code> */ boolean hasSignatureKeyId(); /** * <code>optional int64 signatureKeyId = 2;</code> */ long getSignatureKeyId(); // optional .QueryIdentifierProto query_identifier = 3; /** * <code>optional .QueryIdentifierProto query_identifier = 3;</code> */ boolean hasQueryIdentifier(); /** * <code>optional .QueryIdentifierProto query_identifier = 3;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier(); /** * <code>optional .QueryIdentifierProto query_identifier = 3;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder(); // optional string hive_query_id = 4; /** * <code>optional string hive_query_id = 4;</code> */ boolean hasHiveQueryId(); /** * <code>optional string hive_query_id = 4;</code> */ java.lang.String getHiveQueryId(); /** * <code>optional string hive_query_id = 4;</code> */ com.google.protobuf.ByteString getHiveQueryIdBytes(); // optional string dag_name = 5; /** * <code>optional string dag_name = 5;</code> * * <pre> * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here. * </pre> */ boolean hasDagName(); /** * <code>optional string dag_name = 5;</code> * * <pre> * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here. * </pre> */ java.lang.String getDagName(); /** * <code>optional string dag_name = 5;</code> * * <pre> * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here. * </pre> */ com.google.protobuf.ByteString getDagNameBytes(); // optional string vertex_name = 6; /** * <code>optional string vertex_name = 6;</code> */ boolean hasVertexName(); /** * <code>optional string vertex_name = 6;</code> */ java.lang.String getVertexName(); /** * <code>optional string vertex_name = 6;</code> */ com.google.protobuf.ByteString getVertexNameBytes(); // optional int32 vertex_index = 7; /** * <code>optional int32 vertex_index = 7;</code> */ boolean hasVertexIndex(); /** * <code>optional int32 vertex_index = 7;</code> */ int getVertexIndex(); // optional string token_identifier = 8; /** * <code>optional string token_identifier = 8;</code> * * <pre> * The core vertex stuff * </pre> */ boolean hasTokenIdentifier(); /** * <code>optional string token_identifier = 8;</code> * * <pre> * The core vertex stuff * </pre> */ java.lang.String getTokenIdentifier(); /** * <code>optional string token_identifier = 8;</code> * * <pre> * The core vertex stuff * </pre> */ com.google.protobuf.ByteString getTokenIdentifierBytes(); // optional .EntityDescriptorProto processor_descriptor = 9; /** * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code> */ boolean hasProcessorDescriptor(); /** * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getProcessorDescriptor(); /** * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getProcessorDescriptorOrBuilder(); // repeated .IOSpecProto input_specs = 10; /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> getInputSpecsList(); /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getInputSpecs(int index); /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ int getInputSpecsCount(); /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ java.util.List<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> getInputSpecsOrBuilderList(); /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getInputSpecsOrBuilder( int index); // repeated .IOSpecProto output_specs = 11; /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> getOutputSpecsList(); /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getOutputSpecs(int index); /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ int getOutputSpecsCount(); /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ java.util.List<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> getOutputSpecsOrBuilderList(); /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getOutputSpecsOrBuilder( int index); // repeated .GroupInputSpecProto grouped_input_specs = 12; /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto> getGroupedInputSpecsList(); /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getGroupedInputSpecs(int index); /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ int getGroupedInputSpecsCount(); /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ java.util.List<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder> getGroupedInputSpecsOrBuilderList(); /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder getGroupedInputSpecsOrBuilder( int index); // optional int32 vertex_parallelism = 13; /** * <code>optional int32 vertex_parallelism = 13;</code> * * <pre> * An internal field required for Tez. * </pre> */ boolean hasVertexParallelism(); /** * <code>optional int32 vertex_parallelism = 13;</code> * * <pre> * An internal field required for Tez. * </pre> */ int getVertexParallelism(); // optional bool is_external_submission = 14 [default = false]; /** * <code>optional bool is_external_submission = 14 [default = false];</code> */ boolean hasIsExternalSubmission(); /** * <code>optional bool is_external_submission = 14 [default = false];</code> */ boolean getIsExternalSubmission(); } /** * Protobuf type {@code SignableVertexSpec} * * <pre> * The part of SubmitWork that can be signed * </pre> */ public static final class SignableVertexSpec extends com.google.protobuf.GeneratedMessage implements SignableVertexSpecOrBuilder { // Use SignableVertexSpec.newBuilder() to construct. private SignableVertexSpec(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private SignableVertexSpec(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final SignableVertexSpec defaultInstance; public static SignableVertexSpec getDefaultInstance() { return defaultInstance; } public SignableVertexSpec getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SignableVertexSpec( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { bitField0_ |= 0x00000001; user_ = input.readBytes(); break; } case 16: { bitField0_ |= 0x00000002; signatureKeyId_ = input.readInt64(); break; } case 26: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder subBuilder = null; if (((bitField0_ & 0x00000004) == 0x00000004)) { subBuilder = queryIdentifier_.toBuilder(); } queryIdentifier_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(queryIdentifier_); queryIdentifier_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000004; break; } case 34: { bitField0_ |= 0x00000008; hiveQueryId_ = input.readBytes(); break; } case 42: { bitField0_ |= 0x00000010; dagName_ = input.readBytes(); break; } case 50: { bitField0_ |= 0x00000020; vertexName_ = input.readBytes(); break; } case 56: { bitField0_ |= 0x00000040; vertexIndex_ = input.readInt32(); break; } case 66: { bitField0_ |= 0x00000080; tokenIdentifier_ = input.readBytes(); break; } case 74: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder subBuilder = null; if (((bitField0_ & 0x00000100) == 0x00000100)) { subBuilder = processorDescriptor_.toBuilder(); } processorDescriptor_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(processorDescriptor_); processorDescriptor_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000100; break; } case 82: { if (!((mutable_bitField0_ & 0x00000200) == 0x00000200)) { inputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto>(); mutable_bitField0_ |= 0x00000200; } inputSpecs_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.PARSER, extensionRegistry)); break; } case 90: { if (!((mutable_bitField0_ & 0x00000400) == 0x00000400)) { outputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto>(); mutable_bitField0_ |= 0x00000400; } outputSpecs_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.PARSER, extensionRegistry)); break; } case 98: { if (!((mutable_bitField0_ & 0x00000800) == 0x00000800)) { groupedInputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto>(); mutable_bitField0_ |= 0x00000800; } groupedInputSpecs_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.PARSER, extensionRegistry)); break; } case 104: { bitField0_ |= 0x00000200; vertexParallelism_ = input.readInt32(); break; } case 112: { bitField0_ |= 0x00000400; isExternalSubmission_ = input.readBool(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000200) == 0x00000200)) { inputSpecs_ = java.util.Collections.unmodifiableList(inputSpecs_); } if (((mutable_bitField0_ & 0x00000400) == 0x00000400)) { outputSpecs_ = java.util.Collections.unmodifiableList(outputSpecs_); } if (((mutable_bitField0_ & 0x00000800) == 0x00000800)) { groupedInputSpecs_ = java.util.Collections.unmodifiableList(groupedInputSpecs_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder.class); } public static com.google.protobuf.Parser<SignableVertexSpec> PARSER = new com.google.protobuf.AbstractParser<SignableVertexSpec>() { public SignableVertexSpec parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new SignableVertexSpec(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<SignableVertexSpec> getParserForType() { return PARSER; } private int bitField0_; // optional string user = 1; public static final int USER_FIELD_NUMBER = 1; private java.lang.Object user_; /** * <code>optional string user = 1;</code> */ public boolean hasUser() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional string user = 1;</code> */ public java.lang.String getUser() { java.lang.Object ref = user_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { user_ = s; } return s; } } /** * <code>optional string user = 1;</code> */ public com.google.protobuf.ByteString getUserBytes() { java.lang.Object ref = user_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); user_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } // optional int64 signatureKeyId = 2; public static final int SIGNATUREKEYID_FIELD_NUMBER = 2; private long signatureKeyId_; /** * <code>optional int64 signatureKeyId = 2;</code> */ public boolean hasSignatureKeyId() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional int64 signatureKeyId = 2;</code> */ public long getSignatureKeyId() { return signatureKeyId_; } // optional .QueryIdentifierProto query_identifier = 3; public static final int QUERY_IDENTIFIER_FIELD_NUMBER = 3; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_; /** * <code>optional .QueryIdentifierProto query_identifier = 3;</code> */ public boolean hasQueryIdentifier() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional .QueryIdentifierProto query_identifier = 3;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { return queryIdentifier_; } /** * <code>optional .QueryIdentifierProto query_identifier = 3;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() { return queryIdentifier_; } // optional string hive_query_id = 4; public static final int HIVE_QUERY_ID_FIELD_NUMBER = 4; private java.lang.Object hiveQueryId_; /** * <code>optional string hive_query_id = 4;</code> */ public boolean hasHiveQueryId() { return ((bitField0_ & 0x00000008) == 0x00000008); } /** * <code>optional string hive_query_id = 4;</code> */ public java.lang.String getHiveQueryId() { java.lang.Object ref = hiveQueryId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { hiveQueryId_ = s; } return s; } } /** * <code>optional string hive_query_id = 4;</code> */ public com.google.protobuf.ByteString getHiveQueryIdBytes() { java.lang.Object ref = hiveQueryId_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); hiveQueryId_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } // optional string dag_name = 5; public static final int DAG_NAME_FIELD_NUMBER = 5; private java.lang.Object dagName_; /** * <code>optional string dag_name = 5;</code> * * <pre> * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here. * </pre> */ public boolean hasDagName() { return ((bitField0_ & 0x00000010) == 0x00000010); } /** * <code>optional string dag_name = 5;</code> * * <pre> * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here. * </pre> */ public java.lang.String getDagName() { java.lang.Object ref = dagName_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { dagName_ = s; } return s; } } /** * <code>optional string dag_name = 5;</code> * * <pre> * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here. * </pre> */ public com.google.protobuf.ByteString getDagNameBytes() { java.lang.Object ref = dagName_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); dagName_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } // optional string vertex_name = 6; public static final int VERTEX_NAME_FIELD_NUMBER = 6; private java.lang.Object vertexName_; /** * <code>optional string vertex_name = 6;</code> */ public boolean hasVertexName() { return ((bitField0_ & 0x00000020) == 0x00000020); } /** * <code>optional string vertex_name = 6;</code> */ public java.lang.String getVertexName() { java.lang.Object ref = vertexName_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { vertexName_ = s; } return s; } } /** * <code>optional string vertex_name = 6;</code> */ public com.google.protobuf.ByteString getVertexNameBytes() { java.lang.Object ref = vertexName_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); vertexName_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } // optional int32 vertex_index = 7; public static final int VERTEX_INDEX_FIELD_NUMBER = 7; private int vertexIndex_; /** * <code>optional int32 vertex_index = 7;</code> */ public boolean hasVertexIndex() { return ((bitField0_ & 0x00000040) == 0x00000040); } /** * <code>optional int32 vertex_index = 7;</code> */ public int getVertexIndex() { return vertexIndex_; } // optional string token_identifier = 8; public static final int TOKEN_IDENTIFIER_FIELD_NUMBER = 8; private java.lang.Object tokenIdentifier_; /** * <code>optional string token_identifier = 8;</code> * * <pre> * The core vertex stuff * </pre> */ public boolean hasTokenIdentifier() { return ((bitField0_ & 0x00000080) == 0x00000080); } /** * <code>optional string token_identifier = 8;</code> * * <pre> * The core vertex stuff * </pre> */ public java.lang.String getTokenIdentifier() { java.lang.Object ref = tokenIdentifier_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { tokenIdentifier_ = s; } return s; } } /** * <code>optional string token_identifier = 8;</code> * * <pre> * The core vertex stuff * </pre> */ public com.google.protobuf.ByteString getTokenIdentifierBytes() { java.lang.Object ref = tokenIdentifier_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); tokenIdentifier_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } // optional .EntityDescriptorProto processor_descriptor = 9; public static final int PROCESSOR_DESCRIPTOR_FIELD_NUMBER = 9; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto processorDescriptor_; /** * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code> */ public boolean hasProcessorDescriptor() { return ((bitField0_ & 0x00000100) == 0x00000100); } /** * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getProcessorDescriptor() { return processorDescriptor_; } /** * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getProcessorDescriptorOrBuilder() { return processorDescriptor_; } // repeated .IOSpecProto input_specs = 10; public static final int INPUT_SPECS_FIELD_NUMBER = 10; private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> inputSpecs_; /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> getInputSpecsList() { return inputSpecs_; } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public java.util.List<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> getInputSpecsOrBuilderList() { return inputSpecs_; } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public int getInputSpecsCount() { return inputSpecs_.size(); } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getInputSpecs(int index) { return inputSpecs_.get(index); } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getInputSpecsOrBuilder( int index) { return inputSpecs_.get(index); } // repeated .IOSpecProto output_specs = 11; public static final int OUTPUT_SPECS_FIELD_NUMBER = 11; private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> outputSpecs_; /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> getOutputSpecsList() { return outputSpecs_; } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public java.util.List<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> getOutputSpecsOrBuilderList() { return outputSpecs_; } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public int getOutputSpecsCount() { return outputSpecs_.size(); } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getOutputSpecs(int index) { return outputSpecs_.get(index); } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getOutputSpecsOrBuilder( int index) { return outputSpecs_.get(index); } // repeated .GroupInputSpecProto grouped_input_specs = 12; public static final int GROUPED_INPUT_SPECS_FIELD_NUMBER = 12; private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto> groupedInputSpecs_; /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto> getGroupedInputSpecsList() { return groupedInputSpecs_; } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public java.util.List<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder> getGroupedInputSpecsOrBuilderList() { return groupedInputSpecs_; } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public int getGroupedInputSpecsCount() { return groupedInputSpecs_.size(); } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getGroupedInputSpecs(int index) { return groupedInputSpecs_.get(index); } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder getGroupedInputSpecsOrBuilder( int index) { return groupedInputSpecs_.get(index); } // optional int32 vertex_parallelism = 13; public static final int VERTEX_PARALLELISM_FIELD_NUMBER = 13; private int vertexParallelism_; /** * <code>optional int32 vertex_parallelism = 13;</code> * * <pre> * An internal field required for Tez. * </pre> */ public boolean hasVertexParallelism() { return ((bitField0_ & 0x00000200) == 0x00000200); } /** * <code>optional int32 vertex_parallelism = 13;</code> * * <pre> * An internal field required for Tez. * </pre> */ public int getVertexParallelism() { return vertexParallelism_; } // optional bool is_external_submission = 14 [default = false]; public static final int IS_EXTERNAL_SUBMISSION_FIELD_NUMBER = 14; private boolean isExternalSubmission_; /** * <code>optional bool is_external_submission = 14 [default = false];</code> */ public boolean hasIsExternalSubmission() { return ((bitField0_ & 0x00000400) == 0x00000400); } /** * <code>optional bool is_external_submission = 14 [default = false];</code> */ public boolean getIsExternalSubmission() { return isExternalSubmission_; } private void initFields() { user_ = ""; signatureKeyId_ = 0L; queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); hiveQueryId_ = ""; dagName_ = ""; vertexName_ = ""; vertexIndex_ = 0; tokenIdentifier_ = ""; processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); inputSpecs_ = java.util.Collections.emptyList(); outputSpecs_ = java.util.Collections.emptyList(); groupedInputSpecs_ = java.util.Collections.emptyList(); vertexParallelism_ = 0; isExternalSubmission_ = false; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeBytes(1, getUserBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeInt64(2, signatureKeyId_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { output.writeMessage(3, queryIdentifier_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { output.writeBytes(4, getHiveQueryIdBytes()); } if (((bitField0_ & 0x00000010) == 0x00000010)) { output.writeBytes(5, getDagNameBytes()); } if (((bitField0_ & 0x00000020) == 0x00000020)) { output.writeBytes(6, getVertexNameBytes()); } if (((bitField0_ & 0x00000040) == 0x00000040)) { output.writeInt32(7, vertexIndex_); } if (((bitField0_ & 0x00000080) == 0x00000080)) { output.writeBytes(8, getTokenIdentifierBytes()); } if (((bitField0_ & 0x00000100) == 0x00000100)) { output.writeMessage(9, processorDescriptor_); } for (int i = 0; i < inputSpecs_.size(); i++) { output.writeMessage(10, inputSpecs_.get(i)); } for (int i = 0; i < outputSpecs_.size(); i++) { output.writeMessage(11, outputSpecs_.get(i)); } for (int i = 0; i < groupedInputSpecs_.size(); i++) { output.writeMessage(12, groupedInputSpecs_.get(i)); } if (((bitField0_ & 0x00000200) == 0x00000200)) { output.writeInt32(13, vertexParallelism_); } if (((bitField0_ & 0x00000400) == 0x00000400)) { output.writeBool(14, isExternalSubmission_); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, getUserBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(2, signatureKeyId_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(3, queryIdentifier_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(4, getHiveQueryIdBytes()); } if (((bitField0_ & 0x00000010) == 0x00000010)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(5, getDagNameBytes()); } if (((bitField0_ & 0x00000020) == 0x00000020)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(6, getVertexNameBytes()); } if (((bitField0_ & 0x00000040) == 0x00000040)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(7, vertexIndex_); } if (((bitField0_ & 0x00000080) == 0x00000080)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(8, getTokenIdentifierBytes()); } if (((bitField0_ & 0x00000100) == 0x00000100)) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(9, processorDescriptor_); } for (int i = 0; i < inputSpecs_.size(); i++) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(10, inputSpecs_.get(i)); } for (int i = 0; i < outputSpecs_.size(); i++) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(11, outputSpecs_.get(i)); } for (int i = 0; i < groupedInputSpecs_.size(); i++) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(12, groupedInputSpecs_.get(i)); } if (((bitField0_ & 0x00000200) == 0x00000200)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(13, vertexParallelism_); } if (((bitField0_ & 0x00000400) == 0x00000400)) { size += com.google.protobuf.CodedOutputStream .computeBoolSize(14, isExternalSubmission_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec) obj; boolean result = true; result = result && (hasUser() == other.hasUser()); if (hasUser()) { result = result && getUser() .equals(other.getUser()); } result = result && (hasSignatureKeyId() == other.hasSignatureKeyId()); if (hasSignatureKeyId()) { result = result && (getSignatureKeyId() == other.getSignatureKeyId()); } result = result && (hasQueryIdentifier() == other.hasQueryIdentifier()); if (hasQueryIdentifier()) { result = result && getQueryIdentifier() .equals(other.getQueryIdentifier()); } result = result && (hasHiveQueryId() == other.hasHiveQueryId()); if (hasHiveQueryId()) { result = result && getHiveQueryId() .equals(other.getHiveQueryId()); } result = result && (hasDagName() == other.hasDagName()); if (hasDagName()) { result = result && getDagName() .equals(other.getDagName()); } result = result && (hasVertexName() == other.hasVertexName()); if (hasVertexName()) { result = result && getVertexName() .equals(other.getVertexName()); } result = result && (hasVertexIndex() == other.hasVertexIndex()); if (hasVertexIndex()) { result = result && (getVertexIndex() == other.getVertexIndex()); } result = result && (hasTokenIdentifier() == other.hasTokenIdentifier()); if (hasTokenIdentifier()) { result = result && getTokenIdentifier() .equals(other.getTokenIdentifier()); } result = result && (hasProcessorDescriptor() == other.hasProcessorDescriptor()); if (hasProcessorDescriptor()) { result = result && getProcessorDescriptor() .equals(other.getProcessorDescriptor()); } result = result && getInputSpecsList() .equals(other.getInputSpecsList()); result = result && getOutputSpecsList() .equals(other.getOutputSpecsList()); result = result && getGroupedInputSpecsList() .equals(other.getGroupedInputSpecsList()); result = result && (hasVertexParallelism() == other.hasVertexParallelism()); if (hasVertexParallelism()) { result = result && (getVertexParallelism() == other.getVertexParallelism()); } result = result && (hasIsExternalSubmission() == other.hasIsExternalSubmission()); if (hasIsExternalSubmission()) { result = result && (getIsExternalSubmission() == other.getIsExternalSubmission()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasUser()) { hash = (37 * hash) + USER_FIELD_NUMBER; hash = (53 * hash) + getUser().hashCode(); } if (hasSignatureKeyId()) { hash = (37 * hash) + SIGNATUREKEYID_FIELD_NUMBER; hash = (53 * hash) + hashLong(getSignatureKeyId()); } if (hasQueryIdentifier()) { hash = (37 * hash) + QUERY_IDENTIFIER_FIELD_NUMBER; hash = (53 * hash) + getQueryIdentifier().hashCode(); } if (hasHiveQueryId()) { hash = (37 * hash) + HIVE_QUERY_ID_FIELD_NUMBER; hash = (53 * hash) + getHiveQueryId().hashCode(); } if (hasDagName()) { hash = (37 * hash) + DAG_NAME_FIELD_NUMBER; hash = (53 * hash) + getDagName().hashCode(); } if (hasVertexName()) { hash = (37 * hash) + VERTEX_NAME_FIELD_NUMBER; hash = (53 * hash) + getVertexName().hashCode(); } if (hasVertexIndex()) { hash = (37 * hash) + VERTEX_INDEX_FIELD_NUMBER; hash = (53 * hash) + getVertexIndex(); } if (hasTokenIdentifier()) { hash = (37 * hash) + TOKEN_IDENTIFIER_FIELD_NUMBER; hash = (53 * hash) + getTokenIdentifier().hashCode(); } if (hasProcessorDescriptor()) { hash = (37 * hash) + PROCESSOR_DESCRIPTOR_FIELD_NUMBER; hash = (53 * hash) + getProcessorDescriptor().hashCode(); } if (getInputSpecsCount() > 0) { hash = (37 * hash) + INPUT_SPECS_FIELD_NUMBER; hash = (53 * hash) + getInputSpecsList().hashCode(); } if (getOutputSpecsCount() > 0) { hash = (37 * hash) + OUTPUT_SPECS_FIELD_NUMBER; hash = (53 * hash) + getOutputSpecsList().hashCode(); } if (getGroupedInputSpecsCount() > 0) { hash = (37 * hash) + GROUPED_INPUT_SPECS_FIELD_NUMBER; hash = (53 * hash) + getGroupedInputSpecsList().hashCode(); } if (hasVertexParallelism()) { hash = (37 * hash) + VERTEX_PARALLELISM_FIELD_NUMBER; hash = (53 * hash) + getVertexParallelism(); } if (hasIsExternalSubmission()) { hash = (37 * hash) + IS_EXTERNAL_SUBMISSION_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getIsExternalSubmission()); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code SignableVertexSpec} * * <pre> * The part of SubmitWork that can be signed * </pre> */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getQueryIdentifierFieldBuilder(); getProcessorDescriptorFieldBuilder(); getInputSpecsFieldBuilder(); getOutputSpecsFieldBuilder(); getGroupedInputSpecsFieldBuilder(); } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); user_ = ""; bitField0_ = (bitField0_ & ~0x00000001); signatureKeyId_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); if (queryIdentifierBuilder_ == null) { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); } else { queryIdentifierBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000004); hiveQueryId_ = ""; bitField0_ = (bitField0_ & ~0x00000008); dagName_ = ""; bitField0_ = (bitField0_ & ~0x00000010); vertexName_ = ""; bitField0_ = (bitField0_ & ~0x00000020); vertexIndex_ = 0; bitField0_ = (bitField0_ & ~0x00000040); tokenIdentifier_ = ""; bitField0_ = (bitField0_ & ~0x00000080); if (processorDescriptorBuilder_ == null) { processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); } else { processorDescriptorBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000100); if (inputSpecsBuilder_ == null) { inputSpecs_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000200); } else { inputSpecsBuilder_.clear(); } if (outputSpecsBuilder_ == null) { outputSpecs_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000400); } else { outputSpecsBuilder_.clear(); } if (groupedInputSpecsBuilder_ == null) { groupedInputSpecs_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000800); } else { groupedInputSpecsBuilder_.clear(); } vertexParallelism_ = 0; bitField0_ = (bitField0_ & ~0x00001000); isExternalSubmission_ = false; bitField0_ = (bitField0_ & ~0x00002000); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.user_ = user_; if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } result.signatureKeyId_ = signatureKeyId_; if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000004; } if (queryIdentifierBuilder_ == null) { result.queryIdentifier_ = queryIdentifier_; } else { result.queryIdentifier_ = queryIdentifierBuilder_.build(); } if (((from_bitField0_ & 0x00000008) == 0x00000008)) { to_bitField0_ |= 0x00000008; } result.hiveQueryId_ = hiveQueryId_; if (((from_bitField0_ & 0x00000010) == 0x00000010)) { to_bitField0_ |= 0x00000010; } result.dagName_ = dagName_; if (((from_bitField0_ & 0x00000020) == 0x00000020)) { to_bitField0_ |= 0x00000020; } result.vertexName_ = vertexName_; if (((from_bitField0_ & 0x00000040) == 0x00000040)) { to_bitField0_ |= 0x00000040; } result.vertexIndex_ = vertexIndex_; if (((from_bitField0_ & 0x00000080) == 0x00000080)) { to_bitField0_ |= 0x00000080; } result.tokenIdentifier_ = tokenIdentifier_; if (((from_bitField0_ & 0x00000100) == 0x00000100)) { to_bitField0_ |= 0x00000100; } if (processorDescriptorBuilder_ == null) { result.processorDescriptor_ = processorDescriptor_; } else { result.processorDescriptor_ = processorDescriptorBuilder_.build(); } if (inputSpecsBuilder_ == null) { if (((bitField0_ & 0x00000200) == 0x00000200)) { inputSpecs_ = java.util.Collections.unmodifiableList(inputSpecs_); bitField0_ = (bitField0_ & ~0x00000200); } result.inputSpecs_ = inputSpecs_; } else { result.inputSpecs_ = inputSpecsBuilder_.build(); } if (outputSpecsBuilder_ == null) { if (((bitField0_ & 0x00000400) == 0x00000400)) { outputSpecs_ = java.util.Collections.unmodifiableList(outputSpecs_); bitField0_ = (bitField0_ & ~0x00000400); } result.outputSpecs_ = outputSpecs_; } else { result.outputSpecs_ = outputSpecsBuilder_.build(); } if (groupedInputSpecsBuilder_ == null) { if (((bitField0_ & 0x00000800) == 0x00000800)) { groupedInputSpecs_ = java.util.Collections.unmodifiableList(groupedInputSpecs_); bitField0_ = (bitField0_ & ~0x00000800); } result.groupedInputSpecs_ = groupedInputSpecs_; } else { result.groupedInputSpecs_ = groupedInputSpecsBuilder_.build(); } if (((from_bitField0_ & 0x00001000) == 0x00001000)) { to_bitField0_ |= 0x00000200; } result.vertexParallelism_ = vertexParallelism_; if (((from_bitField0_ & 0x00002000) == 0x00002000)) { to_bitField0_ |= 0x00000400; } result.isExternalSubmission_ = isExternalSubmission_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance()) return this; if (other.hasUser()) { bitField0_ |= 0x00000001; user_ = other.user_; onChanged(); } if (other.hasSignatureKeyId()) { setSignatureKeyId(other.getSignatureKeyId()); } if (other.hasQueryIdentifier()) { mergeQueryIdentifier(other.getQueryIdentifier()); } if (other.hasHiveQueryId()) { bitField0_ |= 0x00000008; hiveQueryId_ = other.hiveQueryId_; onChanged(); } if (other.hasDagName()) { bitField0_ |= 0x00000010; dagName_ = other.dagName_; onChanged(); } if (other.hasVertexName()) { bitField0_ |= 0x00000020; vertexName_ = other.vertexName_; onChanged(); } if (other.hasVertexIndex()) { setVertexIndex(other.getVertexIndex()); } if (other.hasTokenIdentifier()) { bitField0_ |= 0x00000080; tokenIdentifier_ = other.tokenIdentifier_; onChanged(); } if (other.hasProcessorDescriptor()) { mergeProcessorDescriptor(other.getProcessorDescriptor()); } if (inputSpecsBuilder_ == null) { if (!other.inputSpecs_.isEmpty()) { if (inputSpecs_.isEmpty()) { inputSpecs_ = other.inputSpecs_; bitField0_ = (bitField0_ & ~0x00000200); } else { ensureInputSpecsIsMutable(); inputSpecs_.addAll(other.inputSpecs_); } onChanged(); } } else { if (!other.inputSpecs_.isEmpty()) { if (inputSpecsBuilder_.isEmpty()) { inputSpecsBuilder_.dispose(); inputSpecsBuilder_ = null; inputSpecs_ = other.inputSpecs_; bitField0_ = (bitField0_ & ~0x00000200); inputSpecsBuilder_ = com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getInputSpecsFieldBuilder() : null; } else { inputSpecsBuilder_.addAllMessages(other.inputSpecs_); } } } if (outputSpecsBuilder_ == null) { if (!other.outputSpecs_.isEmpty()) { if (outputSpecs_.isEmpty()) { outputSpecs_ = other.outputSpecs_; bitField0_ = (bitField0_ & ~0x00000400); } else { ensureOutputSpecsIsMutable(); outputSpecs_.addAll(other.outputSpecs_); } onChanged(); } } else { if (!other.outputSpecs_.isEmpty()) { if (outputSpecsBuilder_.isEmpty()) { outputSpecsBuilder_.dispose(); outputSpecsBuilder_ = null; outputSpecs_ = other.outputSpecs_; bitField0_ = (bitField0_ & ~0x00000400); outputSpecsBuilder_ = com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getOutputSpecsFieldBuilder() : null; } else { outputSpecsBuilder_.addAllMessages(other.outputSpecs_); } } } if (groupedInputSpecsBuilder_ == null) { if (!other.groupedInputSpecs_.isEmpty()) { if (groupedInputSpecs_.isEmpty()) { groupedInputSpecs_ = other.groupedInputSpecs_; bitField0_ = (bitField0_ & ~0x00000800); } else { ensureGroupedInputSpecsIsMutable(); groupedInputSpecs_.addAll(other.groupedInputSpecs_); } onChanged(); } } else { if (!other.groupedInputSpecs_.isEmpty()) { if (groupedInputSpecsBuilder_.isEmpty()) { groupedInputSpecsBuilder_.dispose(); groupedInputSpecsBuilder_ = null; groupedInputSpecs_ = other.groupedInputSpecs_; bitField0_ = (bitField0_ & ~0x00000800); groupedInputSpecsBuilder_ = com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getGroupedInputSpecsFieldBuilder() : null; } else { groupedInputSpecsBuilder_.addAllMessages(other.groupedInputSpecs_); } } } if (other.hasVertexParallelism()) { setVertexParallelism(other.getVertexParallelism()); } if (other.hasIsExternalSubmission()) { setIsExternalSubmission(other.getIsExternalSubmission()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // optional string user = 1; private java.lang.Object user_ = ""; /** * <code>optional string user = 1;</code> */ public boolean hasUser() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional string user = 1;</code> */ public java.lang.String getUser() { java.lang.Object ref = user_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); user_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>optional string user = 1;</code> */ public com.google.protobuf.ByteString getUserBytes() { java.lang.Object ref = user_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); user_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>optional string user = 1;</code> */ public Builder setUser( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; user_ = value; onChanged(); return this; } /** * <code>optional string user = 1;</code> */ public Builder clearUser() { bitField0_ = (bitField0_ & ~0x00000001); user_ = getDefaultInstance().getUser(); onChanged(); return this; } /** * <code>optional string user = 1;</code> */ public Builder setUserBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; user_ = value; onChanged(); return this; } // optional int64 signatureKeyId = 2; private long signatureKeyId_ ; /** * <code>optional int64 signatureKeyId = 2;</code> */ public boolean hasSignatureKeyId() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional int64 signatureKeyId = 2;</code> */ public long getSignatureKeyId() { return signatureKeyId_; } /** * <code>optional int64 signatureKeyId = 2;</code> */ public Builder setSignatureKeyId(long value) { bitField0_ |= 0x00000002; signatureKeyId_ = value; onChanged(); return this; } /** * <code>optional int64 signatureKeyId = 2;</code> */ public Builder clearSignatureKeyId() { bitField0_ = (bitField0_ & ~0x00000002); signatureKeyId_ = 0L; onChanged(); return this; } // optional .QueryIdentifierProto query_identifier = 3; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> queryIdentifierBuilder_; /** * <code>optional .QueryIdentifierProto query_identifier = 3;</code> */ public boolean hasQueryIdentifier() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional .QueryIdentifierProto query_identifier = 3;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { if (queryIdentifierBuilder_ == null) { return queryIdentifier_; } else { return queryIdentifierBuilder_.getMessage(); } } /** * <code>optional .QueryIdentifierProto query_identifier = 3;</code> */ public Builder setQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto value) { if (queryIdentifierBuilder_ == null) { if (value == null) { throw new NullPointerException(); } queryIdentifier_ = value; onChanged(); } else { queryIdentifierBuilder_.setMessage(value); } bitField0_ |= 0x00000004; return this; } /** * <code>optional .QueryIdentifierProto query_identifier = 3;</code> */ public Builder setQueryIdentifier( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder builderForValue) { if (queryIdentifierBuilder_ == null) { queryIdentifier_ = builderForValue.build(); onChanged(); } else { queryIdentifierBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000004; return this; } /** * <code>optional .QueryIdentifierProto query_identifier = 3;</code> */ public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto value) { if (queryIdentifierBuilder_ == null) { if (((bitField0_ & 0x00000004) == 0x00000004) && queryIdentifier_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance()) { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.newBuilder(queryIdentifier_).mergeFrom(value).buildPartial(); } else { queryIdentifier_ = value; } onChanged(); } else { queryIdentifierBuilder_.mergeFrom(value); } bitField0_ |= 0x00000004; return this; } /** * <code>optional .QueryIdentifierProto query_identifier = 3;</code> */ public Builder clearQueryIdentifier() { if (queryIdentifierBuilder_ == null) { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); onChanged(); } else { queryIdentifierBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000004); return this; } /** * <code>optional .QueryIdentifierProto query_identifier = 3;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder getQueryIdentifierBuilder() { bitField0_ |= 0x00000004; onChanged(); return getQueryIdentifierFieldBuilder().getBuilder(); } /** * <code>optional .QueryIdentifierProto query_identifier = 3;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() { if (queryIdentifierBuilder_ != null) { return queryIdentifierBuilder_.getMessageOrBuilder(); } else { return queryIdentifier_; } } /** * <code>optional .QueryIdentifierProto query_identifier = 3;</code> */ private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> getQueryIdentifierFieldBuilder() { if (queryIdentifierBuilder_ == null) { queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder>( queryIdentifier_, getParentForChildren(), isClean()); queryIdentifier_ = null; } return queryIdentifierBuilder_; } // optional string hive_query_id = 4; private java.lang.Object hiveQueryId_ = ""; /** * <code>optional string hive_query_id = 4;</code> */ public boolean hasHiveQueryId() { return ((bitField0_ & 0x00000008) == 0x00000008); } /** * <code>optional string hive_query_id = 4;</code> */ public java.lang.String getHiveQueryId() { java.lang.Object ref = hiveQueryId_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); hiveQueryId_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>optional string hive_query_id = 4;</code> */ public com.google.protobuf.ByteString getHiveQueryIdBytes() { java.lang.Object ref = hiveQueryId_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); hiveQueryId_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>optional string hive_query_id = 4;</code> */ public Builder setHiveQueryId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000008; hiveQueryId_ = value; onChanged(); return this; } /** * <code>optional string hive_query_id = 4;</code> */ public Builder clearHiveQueryId() { bitField0_ = (bitField0_ & ~0x00000008); hiveQueryId_ = getDefaultInstance().getHiveQueryId(); onChanged(); return this; } /** * <code>optional string hive_query_id = 4;</code> */ public Builder setHiveQueryIdBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000008; hiveQueryId_ = value; onChanged(); return this; } // optional string dag_name = 5; private java.lang.Object dagName_ = ""; /** * <code>optional string dag_name = 5;</code> * * <pre> * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here. * </pre> */ public boolean hasDagName() { return ((bitField0_ & 0x00000010) == 0x00000010); } /** * <code>optional string dag_name = 5;</code> * * <pre> * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here. * </pre> */ public java.lang.String getDagName() { java.lang.Object ref = dagName_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); dagName_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>optional string dag_name = 5;</code> * * <pre> * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here. * </pre> */ public com.google.protobuf.ByteString getDagNameBytes() { java.lang.Object ref = dagName_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); dagName_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>optional string dag_name = 5;</code> * * <pre> * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here. * </pre> */ public Builder setDagName( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000010; dagName_ = value; onChanged(); return this; } /** * <code>optional string dag_name = 5;</code> * * <pre> * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here. * </pre> */ public Builder clearDagName() { bitField0_ = (bitField0_ & ~0x00000010); dagName_ = getDefaultInstance().getDagName(); onChanged(); return this; } /** * <code>optional string dag_name = 5;</code> * * <pre> * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here. * </pre> */ public Builder setDagNameBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000010; dagName_ = value; onChanged(); return this; } // optional string vertex_name = 6; private java.lang.Object vertexName_ = ""; /** * <code>optional string vertex_name = 6;</code> */ public boolean hasVertexName() { return ((bitField0_ & 0x00000020) == 0x00000020); } /** * <code>optional string vertex_name = 6;</code> */ public java.lang.String getVertexName() { java.lang.Object ref = vertexName_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); vertexName_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>optional string vertex_name = 6;</code> */ public com.google.protobuf.ByteString getVertexNameBytes() { java.lang.Object ref = vertexName_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); vertexName_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>optional string vertex_name = 6;</code> */ public Builder setVertexName( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000020; vertexName_ = value; onChanged(); return this; } /** * <code>optional string vertex_name = 6;</code> */ public Builder clearVertexName() { bitField0_ = (bitField0_ & ~0x00000020); vertexName_ = getDefaultInstance().getVertexName(); onChanged(); return this; } /** * <code>optional string vertex_name = 6;</code> */ public Builder setVertexNameBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000020; vertexName_ = value; onChanged(); return this; } // optional int32 vertex_index = 7; private int vertexIndex_ ; /** * <code>optional int32 vertex_index = 7;</code> */ public boolean hasVertexIndex() { return ((bitField0_ & 0x00000040) == 0x00000040); } /** * <code>optional int32 vertex_index = 7;</code> */ public int getVertexIndex() { return vertexIndex_; } /** * <code>optional int32 vertex_index = 7;</code> */ public Builder setVertexIndex(int value) { bitField0_ |= 0x00000040; vertexIndex_ = value; onChanged(); return this; } /** * <code>optional int32 vertex_index = 7;</code> */ public Builder clearVertexIndex() { bitField0_ = (bitField0_ & ~0x00000040); vertexIndex_ = 0; onChanged(); return this; } // optional string token_identifier = 8; private java.lang.Object tokenIdentifier_ = ""; /** * <code>optional string token_identifier = 8;</code> * * <pre> * The core vertex stuff * </pre> */ public boolean hasTokenIdentifier() { return ((bitField0_ & 0x00000080) == 0x00000080); } /** * <code>optional string token_identifier = 8;</code> * * <pre> * The core vertex stuff * </pre> */ public java.lang.String getTokenIdentifier() { java.lang.Object ref = tokenIdentifier_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); tokenIdentifier_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>optional string token_identifier = 8;</code> * * <pre> * The core vertex stuff * </pre> */ public com.google.protobuf.ByteString getTokenIdentifierBytes() { java.lang.Object ref = tokenIdentifier_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); tokenIdentifier_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>optional string token_identifier = 8;</code> * * <pre> * The core vertex stuff * </pre> */ public Builder setTokenIdentifier( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000080; tokenIdentifier_ = value; onChanged(); return this; } /** * <code>optional string token_identifier = 8;</code> * * <pre> * The core vertex stuff * </pre> */ public Builder clearTokenIdentifier() { bitField0_ = (bitField0_ & ~0x00000080); tokenIdentifier_ = getDefaultInstance().getTokenIdentifier(); onChanged(); return this; } /** * <code>optional string token_identifier = 8;</code> * * <pre> * The core vertex stuff * </pre> */ public Builder setTokenIdentifierBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000080; tokenIdentifier_ = value; onChanged(); return this; } // optional .EntityDescriptorProto processor_descriptor = 9; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> processorDescriptorBuilder_; /** * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code> */ public boolean hasProcessorDescriptor() { return ((bitField0_ & 0x00000100) == 0x00000100); } /** * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getProcessorDescriptor() { if (processorDescriptorBuilder_ == null) { return processorDescriptor_; } else { return processorDescriptorBuilder_.getMessage(); } } /** * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code> */ public Builder setProcessorDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) { if (processorDescriptorBuilder_ == null) { if (value == null) { throw new NullPointerException(); } processorDescriptor_ = value; onChanged(); } else { processorDescriptorBuilder_.setMessage(value); } bitField0_ |= 0x00000100; return this; } /** * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code> */ public Builder setProcessorDescriptor( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder builderForValue) { if (processorDescriptorBuilder_ == null) { processorDescriptor_ = builderForValue.build(); onChanged(); } else { processorDescriptorBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000100; return this; } /** * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code> */ public Builder mergeProcessorDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) { if (processorDescriptorBuilder_ == null) { if (((bitField0_ & 0x00000100) == 0x00000100) && processorDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) { processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.newBuilder(processorDescriptor_).mergeFrom(value).buildPartial(); } else { processorDescriptor_ = value; } onChanged(); } else { processorDescriptorBuilder_.mergeFrom(value); } bitField0_ |= 0x00000100; return this; } /** * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code> */ public Builder clearProcessorDescriptor() { if (processorDescriptorBuilder_ == null) { processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); onChanged(); } else { processorDescriptorBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000100); return this; } /** * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder getProcessorDescriptorBuilder() { bitField0_ |= 0x00000100; onChanged(); return getProcessorDescriptorFieldBuilder().getBuilder(); } /** * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getProcessorDescriptorOrBuilder() { if (processorDescriptorBuilder_ != null) { return processorDescriptorBuilder_.getMessageOrBuilder(); } else { return processorDescriptor_; } } /** * <code>optional .EntityDescriptorProto processor_descriptor = 9;</code> */ private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> getProcessorDescriptorFieldBuilder() { if (processorDescriptorBuilder_ == null) { processorDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder>( processorDescriptor_, getParentForChildren(), isClean()); processorDescriptor_ = null; } return processorDescriptorBuilder_; } // repeated .IOSpecProto input_specs = 10; private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> inputSpecs_ = java.util.Collections.emptyList(); private void ensureInputSpecsIsMutable() { if (!((bitField0_ & 0x00000200) == 0x00000200)) { inputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto>(inputSpecs_); bitField0_ |= 0x00000200; } } private com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> inputSpecsBuilder_; /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> getInputSpecsList() { if (inputSpecsBuilder_ == null) { return java.util.Collections.unmodifiableList(inputSpecs_); } else { return inputSpecsBuilder_.getMessageList(); } } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public int getInputSpecsCount() { if (inputSpecsBuilder_ == null) { return inputSpecs_.size(); } else { return inputSpecsBuilder_.getCount(); } } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getInputSpecs(int index) { if (inputSpecsBuilder_ == null) { return inputSpecs_.get(index); } else { return inputSpecsBuilder_.getMessage(index); } } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public Builder setInputSpecs( int index, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto value) { if (inputSpecsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureInputSpecsIsMutable(); inputSpecs_.set(index, value); onChanged(); } else { inputSpecsBuilder_.setMessage(index, value); } return this; } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public Builder setInputSpecs( int index, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder builderForValue) { if (inputSpecsBuilder_ == null) { ensureInputSpecsIsMutable(); inputSpecs_.set(index, builderForValue.build()); onChanged(); } else { inputSpecsBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public Builder addInputSpecs(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto value) { if (inputSpecsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureInputSpecsIsMutable(); inputSpecs_.add(value); onChanged(); } else { inputSpecsBuilder_.addMessage(value); } return this; } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public Builder addInputSpecs( int index, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto value) { if (inputSpecsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureInputSpecsIsMutable(); inputSpecs_.add(index, value); onChanged(); } else { inputSpecsBuilder_.addMessage(index, value); } return this; } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public Builder addInputSpecs( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder builderForValue) { if (inputSpecsBuilder_ == null) { ensureInputSpecsIsMutable(); inputSpecs_.add(builderForValue.build()); onChanged(); } else { inputSpecsBuilder_.addMessage(builderForValue.build()); } return this; } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public Builder addInputSpecs( int index, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder builderForValue) { if (inputSpecsBuilder_ == null) { ensureInputSpecsIsMutable(); inputSpecs_.add(index, builderForValue.build()); onChanged(); } else { inputSpecsBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public Builder addAllInputSpecs( java.lang.Iterable<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> values) { if (inputSpecsBuilder_ == null) { ensureInputSpecsIsMutable(); super.addAll(values, inputSpecs_); onChanged(); } else { inputSpecsBuilder_.addAllMessages(values); } return this; } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public Builder clearInputSpecs() { if (inputSpecsBuilder_ == null) { inputSpecs_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000200); onChanged(); } else { inputSpecsBuilder_.clear(); } return this; } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public Builder removeInputSpecs(int index) { if (inputSpecsBuilder_ == null) { ensureInputSpecsIsMutable(); inputSpecs_.remove(index); onChanged(); } else { inputSpecsBuilder_.remove(index); } return this; } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder getInputSpecsBuilder( int index) { return getInputSpecsFieldBuilder().getBuilder(index); } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getInputSpecsOrBuilder( int index) { if (inputSpecsBuilder_ == null) { return inputSpecs_.get(index); } else { return inputSpecsBuilder_.getMessageOrBuilder(index); } } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public java.util.List<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> getInputSpecsOrBuilderList() { if (inputSpecsBuilder_ != null) { return inputSpecsBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(inputSpecs_); } } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder addInputSpecsBuilder() { return getInputSpecsFieldBuilder().addBuilder( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.getDefaultInstance()); } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder addInputSpecsBuilder( int index) { return getInputSpecsFieldBuilder().addBuilder( index, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.getDefaultInstance()); } /** * <code>repeated .IOSpecProto input_specs = 10;</code> */ public java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder> getInputSpecsBuilderList() { return getInputSpecsFieldBuilder().getBuilderList(); } private com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> getInputSpecsFieldBuilder() { if (inputSpecsBuilder_ == null) { inputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder>( inputSpecs_, ((bitField0_ & 0x00000200) == 0x00000200), getParentForChildren(), isClean()); inputSpecs_ = null; } return inputSpecsBuilder_; } // repeated .IOSpecProto output_specs = 11; private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> outputSpecs_ = java.util.Collections.emptyList(); private void ensureOutputSpecsIsMutable() { if (!((bitField0_ & 0x00000400) == 0x00000400)) { outputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto>(outputSpecs_); bitField0_ |= 0x00000400; } } private com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> outputSpecsBuilder_; /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> getOutputSpecsList() { if (outputSpecsBuilder_ == null) { return java.util.Collections.unmodifiableList(outputSpecs_); } else { return outputSpecsBuilder_.getMessageList(); } } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public int getOutputSpecsCount() { if (outputSpecsBuilder_ == null) { return outputSpecs_.size(); } else { return outputSpecsBuilder_.getCount(); } } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getOutputSpecs(int index) { if (outputSpecsBuilder_ == null) { return outputSpecs_.get(index); } else { return outputSpecsBuilder_.getMessage(index); } } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public Builder setOutputSpecs( int index, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto value) { if (outputSpecsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureOutputSpecsIsMutable(); outputSpecs_.set(index, value); onChanged(); } else { outputSpecsBuilder_.setMessage(index, value); } return this; } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public Builder setOutputSpecs( int index, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder builderForValue) { if (outputSpecsBuilder_ == null) { ensureOutputSpecsIsMutable(); outputSpecs_.set(index, builderForValue.build()); onChanged(); } else { outputSpecsBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public Builder addOutputSpecs(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto value) { if (outputSpecsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureOutputSpecsIsMutable(); outputSpecs_.add(value); onChanged(); } else { outputSpecsBuilder_.addMessage(value); } return this; } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public Builder addOutputSpecs( int index, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto value) { if (outputSpecsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureOutputSpecsIsMutable(); outputSpecs_.add(index, value); onChanged(); } else { outputSpecsBuilder_.addMessage(index, value); } return this; } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public Builder addOutputSpecs( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder builderForValue) { if (outputSpecsBuilder_ == null) { ensureOutputSpecsIsMutable(); outputSpecs_.add(builderForValue.build()); onChanged(); } else { outputSpecsBuilder_.addMessage(builderForValue.build()); } return this; } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public Builder addOutputSpecs( int index, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder builderForValue) { if (outputSpecsBuilder_ == null) { ensureOutputSpecsIsMutable(); outputSpecs_.add(index, builderForValue.build()); onChanged(); } else { outputSpecsBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public Builder addAllOutputSpecs( java.lang.Iterable<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto> values) { if (outputSpecsBuilder_ == null) { ensureOutputSpecsIsMutable(); super.addAll(values, outputSpecs_); onChanged(); } else { outputSpecsBuilder_.addAllMessages(values); } return this; } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public Builder clearOutputSpecs() { if (outputSpecsBuilder_ == null) { outputSpecs_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000400); onChanged(); } else { outputSpecsBuilder_.clear(); } return this; } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public Builder removeOutputSpecs(int index) { if (outputSpecsBuilder_ == null) { ensureOutputSpecsIsMutable(); outputSpecs_.remove(index); onChanged(); } else { outputSpecsBuilder_.remove(index); } return this; } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder getOutputSpecsBuilder( int index) { return getOutputSpecsFieldBuilder().getBuilder(index); } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getOutputSpecsOrBuilder( int index) { if (outputSpecsBuilder_ == null) { return outputSpecs_.get(index); } else { return outputSpecsBuilder_.getMessageOrBuilder(index); } } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public java.util.List<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> getOutputSpecsOrBuilderList() { if (outputSpecsBuilder_ != null) { return outputSpecsBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(outputSpecs_); } } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder addOutputSpecsBuilder() { return getOutputSpecsFieldBuilder().addBuilder( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.getDefaultInstance()); } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder addOutputSpecsBuilder( int index) { return getOutputSpecsFieldBuilder().addBuilder( index, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.getDefaultInstance()); } /** * <code>repeated .IOSpecProto output_specs = 11;</code> */ public java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder> getOutputSpecsBuilderList() { return getOutputSpecsFieldBuilder().getBuilderList(); } private com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> getOutputSpecsFieldBuilder() { if (outputSpecsBuilder_ == null) { outputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder>( outputSpecs_, ((bitField0_ & 0x00000400) == 0x00000400), getParentForChildren(), isClean()); outputSpecs_ = null; } return outputSpecsBuilder_; } // repeated .GroupInputSpecProto grouped_input_specs = 12; private java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto> groupedInputSpecs_ = java.util.Collections.emptyList(); private void ensureGroupedInputSpecsIsMutable() { if (!((bitField0_ & 0x00000800) == 0x00000800)) { groupedInputSpecs_ = new java.util.ArrayList<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto>(groupedInputSpecs_); bitField0_ |= 0x00000800; } } private com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder> groupedInputSpecsBuilder_; /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto> getGroupedInputSpecsList() { if (groupedInputSpecsBuilder_ == null) { return java.util.Collections.unmodifiableList(groupedInputSpecs_); } else { return groupedInputSpecsBuilder_.getMessageList(); } } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public int getGroupedInputSpecsCount() { if (groupedInputSpecsBuilder_ == null) { return groupedInputSpecs_.size(); } else { return groupedInputSpecsBuilder_.getCount(); } } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getGroupedInputSpecs(int index) { if (groupedInputSpecsBuilder_ == null) { return groupedInputSpecs_.get(index); } else { return groupedInputSpecsBuilder_.getMessage(index); } } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public Builder setGroupedInputSpecs( int index, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto value) { if (groupedInputSpecsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureGroupedInputSpecsIsMutable(); groupedInputSpecs_.set(index, value); onChanged(); } else { groupedInputSpecsBuilder_.setMessage(index, value); } return this; } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public Builder setGroupedInputSpecs( int index, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder builderForValue) { if (groupedInputSpecsBuilder_ == null) { ensureGroupedInputSpecsIsMutable(); groupedInputSpecs_.set(index, builderForValue.build()); onChanged(); } else { groupedInputSpecsBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public Builder addGroupedInputSpecs(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto value) { if (groupedInputSpecsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureGroupedInputSpecsIsMutable(); groupedInputSpecs_.add(value); onChanged(); } else { groupedInputSpecsBuilder_.addMessage(value); } return this; } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public Builder addGroupedInputSpecs( int index, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto value) { if (groupedInputSpecsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureGroupedInputSpecsIsMutable(); groupedInputSpecs_.add(index, value); onChanged(); } else { groupedInputSpecsBuilder_.addMessage(index, value); } return this; } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public Builder addGroupedInputSpecs( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder builderForValue) { if (groupedInputSpecsBuilder_ == null) { ensureGroupedInputSpecsIsMutable(); groupedInputSpecs_.add(builderForValue.build()); onChanged(); } else { groupedInputSpecsBuilder_.addMessage(builderForValue.build()); } return this; } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public Builder addGroupedInputSpecs( int index, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder builderForValue) { if (groupedInputSpecsBuilder_ == null) { ensureGroupedInputSpecsIsMutable(); groupedInputSpecs_.add(index, builderForValue.build()); onChanged(); } else { groupedInputSpecsBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public Builder addAllGroupedInputSpecs( java.lang.Iterable<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto> values) { if (groupedInputSpecsBuilder_ == null) { ensureGroupedInputSpecsIsMutable(); super.addAll(values, groupedInputSpecs_); onChanged(); } else { groupedInputSpecsBuilder_.addAllMessages(values); } return this; } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public Builder clearGroupedInputSpecs() { if (groupedInputSpecsBuilder_ == null) { groupedInputSpecs_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000800); onChanged(); } else { groupedInputSpecsBuilder_.clear(); } return this; } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public Builder removeGroupedInputSpecs(int index) { if (groupedInputSpecsBuilder_ == null) { ensureGroupedInputSpecsIsMutable(); groupedInputSpecs_.remove(index); onChanged(); } else { groupedInputSpecsBuilder_.remove(index); } return this; } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder getGroupedInputSpecsBuilder( int index) { return getGroupedInputSpecsFieldBuilder().getBuilder(index); } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder getGroupedInputSpecsOrBuilder( int index) { if (groupedInputSpecsBuilder_ == null) { return groupedInputSpecs_.get(index); } else { return groupedInputSpecsBuilder_.getMessageOrBuilder(index); } } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public java.util.List<? extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder> getGroupedInputSpecsOrBuilderList() { if (groupedInputSpecsBuilder_ != null) { return groupedInputSpecsBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(groupedInputSpecs_); } } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder addGroupedInputSpecsBuilder() { return getGroupedInputSpecsFieldBuilder().addBuilder( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.getDefaultInstance()); } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder addGroupedInputSpecsBuilder( int index) { return getGroupedInputSpecsFieldBuilder().addBuilder( index, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.getDefaultInstance()); } /** * <code>repeated .GroupInputSpecProto grouped_input_specs = 12;</code> */ public java.util.List<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder> getGroupedInputSpecsBuilderList() { return getGroupedInputSpecsFieldBuilder().getBuilderList(); } private com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder> getGroupedInputSpecsFieldBuilder() { if (groupedInputSpecsBuilder_ == null) { groupedInputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder>( groupedInputSpecs_, ((bitField0_ & 0x00000800) == 0x00000800), getParentForChildren(), isClean()); groupedInputSpecs_ = null; } return groupedInputSpecsBuilder_; } // optional int32 vertex_parallelism = 13; private int vertexParallelism_ ; /** * <code>optional int32 vertex_parallelism = 13;</code> * * <pre> * An internal field required for Tez. * </pre> */ public boolean hasVertexParallelism() { return ((bitField0_ & 0x00001000) == 0x00001000); } /** * <code>optional int32 vertex_parallelism = 13;</code> * * <pre> * An internal field required for Tez. * </pre> */ public int getVertexParallelism() { return vertexParallelism_; } /** * <code>optional int32 vertex_parallelism = 13;</code> * * <pre> * An internal field required for Tez. * </pre> */ public Builder setVertexParallelism(int value) { bitField0_ |= 0x00001000; vertexParallelism_ = value; onChanged(); return this; } /** * <code>optional int32 vertex_parallelism = 13;</code> * * <pre> * An internal field required for Tez. * </pre> */ public Builder clearVertexParallelism() { bitField0_ = (bitField0_ & ~0x00001000); vertexParallelism_ = 0; onChanged(); return this; } // optional bool is_external_submission = 14 [default = false]; private boolean isExternalSubmission_ ; /** * <code>optional bool is_external_submission = 14 [default = false];</code> */ public boolean hasIsExternalSubmission() { return ((bitField0_ & 0x00002000) == 0x00002000); } /** * <code>optional bool is_external_submission = 14 [default = false];</code> */ public boolean getIsExternalSubmission() { return isExternalSubmission_; } /** * <code>optional bool is_external_submission = 14 [default = false];</code> */ public Builder setIsExternalSubmission(boolean value) { bitField0_ |= 0x00002000; isExternalSubmission_ = value; onChanged(); return this; } /** * <code>optional bool is_external_submission = 14 [default = false];</code> */ public Builder clearIsExternalSubmission() { bitField0_ = (bitField0_ & ~0x00002000); isExternalSubmission_ = false; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:SignableVertexSpec) } static { defaultInstance = new SignableVertexSpec(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:SignableVertexSpec) } public interface VertexOrBinaryOrBuilder extends com.google.protobuf.MessageOrBuilder { // optional .SignableVertexSpec vertex = 1; /** * <code>optional .SignableVertexSpec vertex = 1;</code> */ boolean hasVertex(); /** * <code>optional .SignableVertexSpec vertex = 1;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getVertex(); /** * <code>optional .SignableVertexSpec vertex = 1;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder getVertexOrBuilder(); // optional bytes vertexBinary = 2; /** * <code>optional bytes vertexBinary = 2;</code> * * <pre> * SignableVertexSpec * </pre> */ boolean hasVertexBinary(); /** * <code>optional bytes vertexBinary = 2;</code> * * <pre> * SignableVertexSpec * </pre> */ com.google.protobuf.ByteString getVertexBinary(); } /** * Protobuf type {@code VertexOrBinary} * * <pre> * Union * </pre> */ public static final class VertexOrBinary extends com.google.protobuf.GeneratedMessage implements VertexOrBinaryOrBuilder { // Use VertexOrBinary.newBuilder() to construct. private VertexOrBinary(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private VertexOrBinary(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final VertexOrBinary defaultInstance; public static VertexOrBinary getDefaultInstance() { return defaultInstance; } public VertexOrBinary getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private VertexOrBinary( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = vertex_.toBuilder(); } vertex_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(vertex_); vertex_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 18: { bitField0_ |= 0x00000002; vertexBinary_ = input.readBytes(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.Builder.class); } public static com.google.protobuf.Parser<VertexOrBinary> PARSER = new com.google.protobuf.AbstractParser<VertexOrBinary>() { public VertexOrBinary parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new VertexOrBinary(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<VertexOrBinary> getParserForType() { return PARSER; } private int bitField0_; // optional .SignableVertexSpec vertex = 1; public static final int VERTEX_FIELD_NUMBER = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec vertex_; /** * <code>optional .SignableVertexSpec vertex = 1;</code> */ public boolean hasVertex() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional .SignableVertexSpec vertex = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getVertex() { return vertex_; } /** * <code>optional .SignableVertexSpec vertex = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder getVertexOrBuilder() { return vertex_; } // optional bytes vertexBinary = 2; public static final int VERTEXBINARY_FIELD_NUMBER = 2; private com.google.protobuf.ByteString vertexBinary_; /** * <code>optional bytes vertexBinary = 2;</code> * * <pre> * SignableVertexSpec * </pre> */ public boolean hasVertexBinary() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional bytes vertexBinary = 2;</code> * * <pre> * SignableVertexSpec * </pre> */ public com.google.protobuf.ByteString getVertexBinary() { return vertexBinary_; } private void initFields() { vertex_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance(); vertexBinary_ = com.google.protobuf.ByteString.EMPTY; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeMessage(1, vertex_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeBytes(2, vertexBinary_); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(1, vertex_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(2, vertexBinary_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary) obj; boolean result = true; result = result && (hasVertex() == other.hasVertex()); if (hasVertex()) { result = result && getVertex() .equals(other.getVertex()); } result = result && (hasVertexBinary() == other.hasVertexBinary()); if (hasVertexBinary()) { result = result && getVertexBinary() .equals(other.getVertexBinary()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasVertex()) { hash = (37 * hash) + VERTEX_FIELD_NUMBER; hash = (53 * hash) + getVertex().hashCode(); } if (hasVertexBinary()) { hash = (37 * hash) + VERTEXBINARY_FIELD_NUMBER; hash = (53 * hash) + getVertexBinary().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code VertexOrBinary} * * <pre> * Union * </pre> */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinaryOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getVertexFieldBuilder(); } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); if (vertexBuilder_ == null) { vertex_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance(); } else { vertexBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); vertexBinary_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } if (vertexBuilder_ == null) { result.vertex_ = vertex_; } else { result.vertex_ = vertexBuilder_.build(); } if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } result.vertexBinary_ = vertexBinary_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance()) return this; if (other.hasVertex()) { mergeVertex(other.getVertex()); } if (other.hasVertexBinary()) { setVertexBinary(other.getVertexBinary()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // optional .SignableVertexSpec vertex = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec vertex_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder> vertexBuilder_; /** * <code>optional .SignableVertexSpec vertex = 1;</code> */ public boolean hasVertex() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional .SignableVertexSpec vertex = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getVertex() { if (vertexBuilder_ == null) { return vertex_; } else { return vertexBuilder_.getMessage(); } } /** * <code>optional .SignableVertexSpec vertex = 1;</code> */ public Builder setVertex(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec value) { if (vertexBuilder_ == null) { if (value == null) { throw new NullPointerException(); } vertex_ = value; onChanged(); } else { vertexBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * <code>optional .SignableVertexSpec vertex = 1;</code> */ public Builder setVertex( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder builderForValue) { if (vertexBuilder_ == null) { vertex_ = builderForValue.build(); onChanged(); } else { vertexBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * <code>optional .SignableVertexSpec vertex = 1;</code> */ public Builder mergeVertex(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec value) { if (vertexBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && vertex_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance()) { vertex_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.newBuilder(vertex_).mergeFrom(value).buildPartial(); } else { vertex_ = value; } onChanged(); } else { vertexBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * <code>optional .SignableVertexSpec vertex = 1;</code> */ public Builder clearVertex() { if (vertexBuilder_ == null) { vertex_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance(); onChanged(); } else { vertexBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * <code>optional .SignableVertexSpec vertex = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder getVertexBuilder() { bitField0_ |= 0x00000001; onChanged(); return getVertexFieldBuilder().getBuilder(); } /** * <code>optional .SignableVertexSpec vertex = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder getVertexOrBuilder() { if (vertexBuilder_ != null) { return vertexBuilder_.getMessageOrBuilder(); } else { return vertex_; } } /** * <code>optional .SignableVertexSpec vertex = 1;</code> */ private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder> getVertexFieldBuilder() { if (vertexBuilder_ == null) { vertexBuilder_ = new com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder>( vertex_, getParentForChildren(), isClean()); vertex_ = null; } return vertexBuilder_; } // optional bytes vertexBinary = 2; private com.google.protobuf.ByteString vertexBinary_ = com.google.protobuf.ByteString.EMPTY; /** * <code>optional bytes vertexBinary = 2;</code> * * <pre> * SignableVertexSpec * </pre> */ public boolean hasVertexBinary() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional bytes vertexBinary = 2;</code> * * <pre> * SignableVertexSpec * </pre> */ public com.google.protobuf.ByteString getVertexBinary() { return vertexBinary_; } /** * <code>optional bytes vertexBinary = 2;</code> * * <pre> * SignableVertexSpec * </pre> */ public Builder setVertexBinary(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; vertexBinary_ = value; onChanged(); return this; } /** * <code>optional bytes vertexBinary = 2;</code> * * <pre> * SignableVertexSpec * </pre> */ public Builder clearVertexBinary() { bitField0_ = (bitField0_ & ~0x00000002); vertexBinary_ = getDefaultInstance().getVertexBinary(); onChanged(); return this; } // @@protoc_insertion_point(builder_scope:VertexOrBinary) } static { defaultInstance = new VertexOrBinary(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:VertexOrBinary) } public interface FragmentRuntimeInfoOrBuilder extends com.google.protobuf.MessageOrBuilder { // optional int32 num_self_and_upstream_tasks = 1; /** * <code>optional int32 num_self_and_upstream_tasks = 1;</code> */ boolean hasNumSelfAndUpstreamTasks(); /** * <code>optional int32 num_self_and_upstream_tasks = 1;</code> */ int getNumSelfAndUpstreamTasks(); // optional int32 num_self_and_upstream_completed_tasks = 2; /** * <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code> */ boolean hasNumSelfAndUpstreamCompletedTasks(); /** * <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code> */ int getNumSelfAndUpstreamCompletedTasks(); // optional int32 within_dag_priority = 3; /** * <code>optional int32 within_dag_priority = 3;</code> */ boolean hasWithinDagPriority(); /** * <code>optional int32 within_dag_priority = 3;</code> */ int getWithinDagPriority(); // optional int64 dag_start_time = 4; /** * <code>optional int64 dag_start_time = 4;</code> */ boolean hasDagStartTime(); /** * <code>optional int64 dag_start_time = 4;</code> */ long getDagStartTime(); // optional int64 first_attempt_start_time = 5; /** * <code>optional int64 first_attempt_start_time = 5;</code> */ boolean hasFirstAttemptStartTime(); /** * <code>optional int64 first_attempt_start_time = 5;</code> */ long getFirstAttemptStartTime(); // optional int64 current_attempt_start_time = 6; /** * <code>optional int64 current_attempt_start_time = 6;</code> */ boolean hasCurrentAttemptStartTime(); /** * <code>optional int64 current_attempt_start_time = 6;</code> */ long getCurrentAttemptStartTime(); } /** * Protobuf type {@code FragmentRuntimeInfo} */ public static final class FragmentRuntimeInfo extends com.google.protobuf.GeneratedMessage implements FragmentRuntimeInfoOrBuilder { // Use FragmentRuntimeInfo.newBuilder() to construct. private FragmentRuntimeInfo(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private FragmentRuntimeInfo(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final FragmentRuntimeInfo defaultInstance; public static FragmentRuntimeInfo getDefaultInstance() { return defaultInstance; } public FragmentRuntimeInfo getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private FragmentRuntimeInfo( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 8: { bitField0_ |= 0x00000001; numSelfAndUpstreamTasks_ = input.readInt32(); break; } case 16: { bitField0_ |= 0x00000002; numSelfAndUpstreamCompletedTasks_ = input.readInt32(); break; } case 24: { bitField0_ |= 0x00000004; withinDagPriority_ = input.readInt32(); break; } case 32: { bitField0_ |= 0x00000008; dagStartTime_ = input.readInt64(); break; } case 40: { bitField0_ |= 0x00000010; firstAttemptStartTime_ = input.readInt64(); break; } case 48: { bitField0_ |= 0x00000020; currentAttemptStartTime_ = input.readInt64(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.Builder.class); } public static com.google.protobuf.Parser<FragmentRuntimeInfo> PARSER = new com.google.protobuf.AbstractParser<FragmentRuntimeInfo>() { public FragmentRuntimeInfo parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new FragmentRuntimeInfo(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<FragmentRuntimeInfo> getParserForType() { return PARSER; } private int bitField0_; // optional int32 num_self_and_upstream_tasks = 1; public static final int NUM_SELF_AND_UPSTREAM_TASKS_FIELD_NUMBER = 1; private int numSelfAndUpstreamTasks_; /** * <code>optional int32 num_self_and_upstream_tasks = 1;</code> */ public boolean hasNumSelfAndUpstreamTasks() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional int32 num_self_and_upstream_tasks = 1;</code> */ public int getNumSelfAndUpstreamTasks() { return numSelfAndUpstreamTasks_; } // optional int32 num_self_and_upstream_completed_tasks = 2; public static final int NUM_SELF_AND_UPSTREAM_COMPLETED_TASKS_FIELD_NUMBER = 2; private int numSelfAndUpstreamCompletedTasks_; /** * <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code> */ public boolean hasNumSelfAndUpstreamCompletedTasks() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code> */ public int getNumSelfAndUpstreamCompletedTasks() { return numSelfAndUpstreamCompletedTasks_; } // optional int32 within_dag_priority = 3; public static final int WITHIN_DAG_PRIORITY_FIELD_NUMBER = 3; private int withinDagPriority_; /** * <code>optional int32 within_dag_priority = 3;</code> */ public boolean hasWithinDagPriority() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional int32 within_dag_priority = 3;</code> */ public int getWithinDagPriority() { return withinDagPriority_; } // optional int64 dag_start_time = 4; public static final int DAG_START_TIME_FIELD_NUMBER = 4; private long dagStartTime_; /** * <code>optional int64 dag_start_time = 4;</code> */ public boolean hasDagStartTime() { return ((bitField0_ & 0x00000008) == 0x00000008); } /** * <code>optional int64 dag_start_time = 4;</code> */ public long getDagStartTime() { return dagStartTime_; } // optional int64 first_attempt_start_time = 5; public static final int FIRST_ATTEMPT_START_TIME_FIELD_NUMBER = 5; private long firstAttemptStartTime_; /** * <code>optional int64 first_attempt_start_time = 5;</code> */ public boolean hasFirstAttemptStartTime() { return ((bitField0_ & 0x00000010) == 0x00000010); } /** * <code>optional int64 first_attempt_start_time = 5;</code> */ public long getFirstAttemptStartTime() { return firstAttemptStartTime_; } // optional int64 current_attempt_start_time = 6; public static final int CURRENT_ATTEMPT_START_TIME_FIELD_NUMBER = 6; private long currentAttemptStartTime_; /** * <code>optional int64 current_attempt_start_time = 6;</code> */ public boolean hasCurrentAttemptStartTime() { return ((bitField0_ & 0x00000020) == 0x00000020); } /** * <code>optional int64 current_attempt_start_time = 6;</code> */ public long getCurrentAttemptStartTime() { return currentAttemptStartTime_; } private void initFields() { numSelfAndUpstreamTasks_ = 0; numSelfAndUpstreamCompletedTasks_ = 0; withinDagPriority_ = 0; dagStartTime_ = 0L; firstAttemptStartTime_ = 0L; currentAttemptStartTime_ = 0L; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeInt32(1, numSelfAndUpstreamTasks_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeInt32(2, numSelfAndUpstreamCompletedTasks_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { output.writeInt32(3, withinDagPriority_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { output.writeInt64(4, dagStartTime_); } if (((bitField0_ & 0x00000010) == 0x00000010)) { output.writeInt64(5, firstAttemptStartTime_); } if (((bitField0_ & 0x00000020) == 0x00000020)) { output.writeInt64(6, currentAttemptStartTime_); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(1, numSelfAndUpstreamTasks_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(2, numSelfAndUpstreamCompletedTasks_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(3, withinDagPriority_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(4, dagStartTime_); } if (((bitField0_ & 0x00000010) == 0x00000010)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(5, firstAttemptStartTime_); } if (((bitField0_ & 0x00000020) == 0x00000020)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(6, currentAttemptStartTime_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo) obj; boolean result = true; result = result && (hasNumSelfAndUpstreamTasks() == other.hasNumSelfAndUpstreamTasks()); if (hasNumSelfAndUpstreamTasks()) { result = result && (getNumSelfAndUpstreamTasks() == other.getNumSelfAndUpstreamTasks()); } result = result && (hasNumSelfAndUpstreamCompletedTasks() == other.hasNumSelfAndUpstreamCompletedTasks()); if (hasNumSelfAndUpstreamCompletedTasks()) { result = result && (getNumSelfAndUpstreamCompletedTasks() == other.getNumSelfAndUpstreamCompletedTasks()); } result = result && (hasWithinDagPriority() == other.hasWithinDagPriority()); if (hasWithinDagPriority()) { result = result && (getWithinDagPriority() == other.getWithinDagPriority()); } result = result && (hasDagStartTime() == other.hasDagStartTime()); if (hasDagStartTime()) { result = result && (getDagStartTime() == other.getDagStartTime()); } result = result && (hasFirstAttemptStartTime() == other.hasFirstAttemptStartTime()); if (hasFirstAttemptStartTime()) { result = result && (getFirstAttemptStartTime() == other.getFirstAttemptStartTime()); } result = result && (hasCurrentAttemptStartTime() == other.hasCurrentAttemptStartTime()); if (hasCurrentAttemptStartTime()) { result = result && (getCurrentAttemptStartTime() == other.getCurrentAttemptStartTime()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasNumSelfAndUpstreamTasks()) { hash = (37 * hash) + NUM_SELF_AND_UPSTREAM_TASKS_FIELD_NUMBER; hash = (53 * hash) + getNumSelfAndUpstreamTasks(); } if (hasNumSelfAndUpstreamCompletedTasks()) { hash = (37 * hash) + NUM_SELF_AND_UPSTREAM_COMPLETED_TASKS_FIELD_NUMBER; hash = (53 * hash) + getNumSelfAndUpstreamCompletedTasks(); } if (hasWithinDagPriority()) { hash = (37 * hash) + WITHIN_DAG_PRIORITY_FIELD_NUMBER; hash = (53 * hash) + getWithinDagPriority(); } if (hasDagStartTime()) { hash = (37 * hash) + DAG_START_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getDagStartTime()); } if (hasFirstAttemptStartTime()) { hash = (37 * hash) + FIRST_ATTEMPT_START_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getFirstAttemptStartTime()); } if (hasCurrentAttemptStartTime()) { hash = (37 * hash) + CURRENT_ATTEMPT_START_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getCurrentAttemptStartTime()); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code FragmentRuntimeInfo} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); numSelfAndUpstreamTasks_ = 0; bitField0_ = (bitField0_ & ~0x00000001); numSelfAndUpstreamCompletedTasks_ = 0; bitField0_ = (bitField0_ & ~0x00000002); withinDagPriority_ = 0; bitField0_ = (bitField0_ & ~0x00000004); dagStartTime_ = 0L; bitField0_ = (bitField0_ & ~0x00000008); firstAttemptStartTime_ = 0L; bitField0_ = (bitField0_ & ~0x00000010); currentAttemptStartTime_ = 0L; bitField0_ = (bitField0_ & ~0x00000020); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.numSelfAndUpstreamTasks_ = numSelfAndUpstreamTasks_; if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } result.numSelfAndUpstreamCompletedTasks_ = numSelfAndUpstreamCompletedTasks_; if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000004; } result.withinDagPriority_ = withinDagPriority_; if (((from_bitField0_ & 0x00000008) == 0x00000008)) { to_bitField0_ |= 0x00000008; } result.dagStartTime_ = dagStartTime_; if (((from_bitField0_ & 0x00000010) == 0x00000010)) { to_bitField0_ |= 0x00000010; } result.firstAttemptStartTime_ = firstAttemptStartTime_; if (((from_bitField0_ & 0x00000020) == 0x00000020)) { to_bitField0_ |= 0x00000020; } result.currentAttemptStartTime_ = currentAttemptStartTime_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance()) return this; if (other.hasNumSelfAndUpstreamTasks()) { setNumSelfAndUpstreamTasks(other.getNumSelfAndUpstreamTasks()); } if (other.hasNumSelfAndUpstreamCompletedTasks()) { setNumSelfAndUpstreamCompletedTasks(other.getNumSelfAndUpstreamCompletedTasks()); } if (other.hasWithinDagPriority()) { setWithinDagPriority(other.getWithinDagPriority()); } if (other.hasDagStartTime()) { setDagStartTime(other.getDagStartTime()); } if (other.hasFirstAttemptStartTime()) { setFirstAttemptStartTime(other.getFirstAttemptStartTime()); } if (other.hasCurrentAttemptStartTime()) { setCurrentAttemptStartTime(other.getCurrentAttemptStartTime()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // optional int32 num_self_and_upstream_tasks = 1; private int numSelfAndUpstreamTasks_ ; /** * <code>optional int32 num_self_and_upstream_tasks = 1;</code> */ public boolean hasNumSelfAndUpstreamTasks() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional int32 num_self_and_upstream_tasks = 1;</code> */ public int getNumSelfAndUpstreamTasks() { return numSelfAndUpstreamTasks_; } /** * <code>optional int32 num_self_and_upstream_tasks = 1;</code> */ public Builder setNumSelfAndUpstreamTasks(int value) { bitField0_ |= 0x00000001; numSelfAndUpstreamTasks_ = value; onChanged(); return this; } /** * <code>optional int32 num_self_and_upstream_tasks = 1;</code> */ public Builder clearNumSelfAndUpstreamTasks() { bitField0_ = (bitField0_ & ~0x00000001); numSelfAndUpstreamTasks_ = 0; onChanged(); return this; } // optional int32 num_self_and_upstream_completed_tasks = 2; private int numSelfAndUpstreamCompletedTasks_ ; /** * <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code> */ public boolean hasNumSelfAndUpstreamCompletedTasks() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code> */ public int getNumSelfAndUpstreamCompletedTasks() { return numSelfAndUpstreamCompletedTasks_; } /** * <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code> */ public Builder setNumSelfAndUpstreamCompletedTasks(int value) { bitField0_ |= 0x00000002; numSelfAndUpstreamCompletedTasks_ = value; onChanged(); return this; } /** * <code>optional int32 num_self_and_upstream_completed_tasks = 2;</code> */ public Builder clearNumSelfAndUpstreamCompletedTasks() { bitField0_ = (bitField0_ & ~0x00000002); numSelfAndUpstreamCompletedTasks_ = 0; onChanged(); return this; } // optional int32 within_dag_priority = 3; private int withinDagPriority_ ; /** * <code>optional int32 within_dag_priority = 3;</code> */ public boolean hasWithinDagPriority() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional int32 within_dag_priority = 3;</code> */ public int getWithinDagPriority() { return withinDagPriority_; } /** * <code>optional int32 within_dag_priority = 3;</code> */ public Builder setWithinDagPriority(int value) { bitField0_ |= 0x00000004; withinDagPriority_ = value; onChanged(); return this; } /** * <code>optional int32 within_dag_priority = 3;</code> */ public Builder clearWithinDagPriority() { bitField0_ = (bitField0_ & ~0x00000004); withinDagPriority_ = 0; onChanged(); return this; } // optional int64 dag_start_time = 4; private long dagStartTime_ ; /** * <code>optional int64 dag_start_time = 4;</code> */ public boolean hasDagStartTime() { return ((bitField0_ & 0x00000008) == 0x00000008); } /** * <code>optional int64 dag_start_time = 4;</code> */ public long getDagStartTime() { return dagStartTime_; } /** * <code>optional int64 dag_start_time = 4;</code> */ public Builder setDagStartTime(long value) { bitField0_ |= 0x00000008; dagStartTime_ = value; onChanged(); return this; } /** * <code>optional int64 dag_start_time = 4;</code> */ public Builder clearDagStartTime() { bitField0_ = (bitField0_ & ~0x00000008); dagStartTime_ = 0L; onChanged(); return this; } // optional int64 first_attempt_start_time = 5; private long firstAttemptStartTime_ ; /** * <code>optional int64 first_attempt_start_time = 5;</code> */ public boolean hasFirstAttemptStartTime() { return ((bitField0_ & 0x00000010) == 0x00000010); } /** * <code>optional int64 first_attempt_start_time = 5;</code> */ public long getFirstAttemptStartTime() { return firstAttemptStartTime_; } /** * <code>optional int64 first_attempt_start_time = 5;</code> */ public Builder setFirstAttemptStartTime(long value) { bitField0_ |= 0x00000010; firstAttemptStartTime_ = value; onChanged(); return this; } /** * <code>optional int64 first_attempt_start_time = 5;</code> */ public Builder clearFirstAttemptStartTime() { bitField0_ = (bitField0_ & ~0x00000010); firstAttemptStartTime_ = 0L; onChanged(); return this; } // optional int64 current_attempt_start_time = 6; private long currentAttemptStartTime_ ; /** * <code>optional int64 current_attempt_start_time = 6;</code> */ public boolean hasCurrentAttemptStartTime() { return ((bitField0_ & 0x00000020) == 0x00000020); } /** * <code>optional int64 current_attempt_start_time = 6;</code> */ public long getCurrentAttemptStartTime() { return currentAttemptStartTime_; } /** * <code>optional int64 current_attempt_start_time = 6;</code> */ public Builder setCurrentAttemptStartTime(long value) { bitField0_ |= 0x00000020; currentAttemptStartTime_ = value; onChanged(); return this; } /** * <code>optional int64 current_attempt_start_time = 6;</code> */ public Builder clearCurrentAttemptStartTime() { bitField0_ = (bitField0_ & ~0x00000020); currentAttemptStartTime_ = 0L; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:FragmentRuntimeInfo) } static { defaultInstance = new FragmentRuntimeInfo(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:FragmentRuntimeInfo) } public interface QueryIdentifierProtoOrBuilder extends com.google.protobuf.MessageOrBuilder { // optional string application_id_string = 1; /** * <code>optional string application_id_string = 1;</code> */ boolean hasApplicationIdString(); /** * <code>optional string application_id_string = 1;</code> */ java.lang.String getApplicationIdString(); /** * <code>optional string application_id_string = 1;</code> */ com.google.protobuf.ByteString getApplicationIdStringBytes(); // optional int32 dag_index = 2; /** * <code>optional int32 dag_index = 2;</code> */ boolean hasDagIndex(); /** * <code>optional int32 dag_index = 2;</code> */ int getDagIndex(); // optional int32 app_attempt_number = 3; /** * <code>optional int32 app_attempt_number = 3;</code> */ boolean hasAppAttemptNumber(); /** * <code>optional int32 app_attempt_number = 3;</code> */ int getAppAttemptNumber(); } /** * Protobuf type {@code QueryIdentifierProto} */ public static final class QueryIdentifierProto extends com.google.protobuf.GeneratedMessage implements QueryIdentifierProtoOrBuilder { // Use QueryIdentifierProto.newBuilder() to construct. private QueryIdentifierProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private QueryIdentifierProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final QueryIdentifierProto defaultInstance; public static QueryIdentifierProto getDefaultInstance() { return defaultInstance; } public QueryIdentifierProto getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private QueryIdentifierProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { bitField0_ |= 0x00000001; applicationIdString_ = input.readBytes(); break; } case 16: { bitField0_ |= 0x00000002; dagIndex_ = input.readInt32(); break; } case 24: { bitField0_ |= 0x00000004; appAttemptNumber_ = input.readInt32(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder.class); } public static com.google.protobuf.Parser<QueryIdentifierProto> PARSER = new com.google.protobuf.AbstractParser<QueryIdentifierProto>() { public QueryIdentifierProto parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new QueryIdentifierProto(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<QueryIdentifierProto> getParserForType() { return PARSER; } private int bitField0_; // optional string application_id_string = 1; public static final int APPLICATION_ID_STRING_FIELD_NUMBER = 1; private java.lang.Object applicationIdString_; /** * <code>optional string application_id_string = 1;</code> */ public boolean hasApplicationIdString() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional string application_id_string = 1;</code> */ public java.lang.String getApplicationIdString() { java.lang.Object ref = applicationIdString_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { applicationIdString_ = s; } return s; } } /** * <code>optional string application_id_string = 1;</code> */ public com.google.protobuf.ByteString getApplicationIdStringBytes() { java.lang.Object ref = applicationIdString_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); applicationIdString_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } // optional int32 dag_index = 2; public static final int DAG_INDEX_FIELD_NUMBER = 2; private int dagIndex_; /** * <code>optional int32 dag_index = 2;</code> */ public boolean hasDagIndex() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional int32 dag_index = 2;</code> */ public int getDagIndex() { return dagIndex_; } // optional int32 app_attempt_number = 3; public static final int APP_ATTEMPT_NUMBER_FIELD_NUMBER = 3; private int appAttemptNumber_; /** * <code>optional int32 app_attempt_number = 3;</code> */ public boolean hasAppAttemptNumber() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional int32 app_attempt_number = 3;</code> */ public int getAppAttemptNumber() { return appAttemptNumber_; } private void initFields() { applicationIdString_ = ""; dagIndex_ = 0; appAttemptNumber_ = 0; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeBytes(1, getApplicationIdStringBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeInt32(2, dagIndex_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { output.writeInt32(3, appAttemptNumber_); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, getApplicationIdStringBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(2, dagIndex_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(3, appAttemptNumber_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto) obj; boolean result = true; result = result && (hasApplicationIdString() == other.hasApplicationIdString()); if (hasApplicationIdString()) { result = result && getApplicationIdString() .equals(other.getApplicationIdString()); } result = result && (hasDagIndex() == other.hasDagIndex()); if (hasDagIndex()) { result = result && (getDagIndex() == other.getDagIndex()); } result = result && (hasAppAttemptNumber() == other.hasAppAttemptNumber()); if (hasAppAttemptNumber()) { result = result && (getAppAttemptNumber() == other.getAppAttemptNumber()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasApplicationIdString()) { hash = (37 * hash) + APPLICATION_ID_STRING_FIELD_NUMBER; hash = (53 * hash) + getApplicationIdString().hashCode(); } if (hasDagIndex()) { hash = (37 * hash) + DAG_INDEX_FIELD_NUMBER; hash = (53 * hash) + getDagIndex(); } if (hasAppAttemptNumber()) { hash = (37 * hash) + APP_ATTEMPT_NUMBER_FIELD_NUMBER; hash = (53 * hash) + getAppAttemptNumber(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code QueryIdentifierProto} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); applicationIdString_ = ""; bitField0_ = (bitField0_ & ~0x00000001); dagIndex_ = 0; bitField0_ = (bitField0_ & ~0x00000002); appAttemptNumber_ = 0; bitField0_ = (bitField0_ & ~0x00000004); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.applicationIdString_ = applicationIdString_; if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } result.dagIndex_ = dagIndex_; if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000004; } result.appAttemptNumber_ = appAttemptNumber_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance()) return this; if (other.hasApplicationIdString()) { bitField0_ |= 0x00000001; applicationIdString_ = other.applicationIdString_; onChanged(); } if (other.hasDagIndex()) { setDagIndex(other.getDagIndex()); } if (other.hasAppAttemptNumber()) { setAppAttemptNumber(other.getAppAttemptNumber()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // optional string application_id_string = 1; private java.lang.Object applicationIdString_ = ""; /** * <code>optional string application_id_string = 1;</code> */ public boolean hasApplicationIdString() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional string application_id_string = 1;</code> */ public java.lang.String getApplicationIdString() { java.lang.Object ref = applicationIdString_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); applicationIdString_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>optional string application_id_string = 1;</code> */ public com.google.protobuf.ByteString getApplicationIdStringBytes() { java.lang.Object ref = applicationIdString_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); applicationIdString_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>optional string application_id_string = 1;</code> */ public Builder setApplicationIdString( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; applicationIdString_ = value; onChanged(); return this; } /** * <code>optional string application_id_string = 1;</code> */ public Builder clearApplicationIdString() { bitField0_ = (bitField0_ & ~0x00000001); applicationIdString_ = getDefaultInstance().getApplicationIdString(); onChanged(); return this; } /** * <code>optional string application_id_string = 1;</code> */ public Builder setApplicationIdStringBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; applicationIdString_ = value; onChanged(); return this; } // optional int32 dag_index = 2; private int dagIndex_ ; /** * <code>optional int32 dag_index = 2;</code> */ public boolean hasDagIndex() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional int32 dag_index = 2;</code> */ public int getDagIndex() { return dagIndex_; } /** * <code>optional int32 dag_index = 2;</code> */ public Builder setDagIndex(int value) { bitField0_ |= 0x00000002; dagIndex_ = value; onChanged(); return this; } /** * <code>optional int32 dag_index = 2;</code> */ public Builder clearDagIndex() { bitField0_ = (bitField0_ & ~0x00000002); dagIndex_ = 0; onChanged(); return this; } // optional int32 app_attempt_number = 3; private int appAttemptNumber_ ; /** * <code>optional int32 app_attempt_number = 3;</code> */ public boolean hasAppAttemptNumber() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional int32 app_attempt_number = 3;</code> */ public int getAppAttemptNumber() { return appAttemptNumber_; } /** * <code>optional int32 app_attempt_number = 3;</code> */ public Builder setAppAttemptNumber(int value) { bitField0_ |= 0x00000004; appAttemptNumber_ = value; onChanged(); return this; } /** * <code>optional int32 app_attempt_number = 3;</code> */ public Builder clearAppAttemptNumber() { bitField0_ = (bitField0_ & ~0x00000004); appAttemptNumber_ = 0; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:QueryIdentifierProto) } static { defaultInstance = new QueryIdentifierProto(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:QueryIdentifierProto) } public interface NotTezEventOrBuilder extends com.google.protobuf.MessageOrBuilder { // required bytes input_event_proto_bytes = 1; /** * <code>required bytes input_event_proto_bytes = 1;</code> */ boolean hasInputEventProtoBytes(); /** * <code>required bytes input_event_proto_bytes = 1;</code> */ com.google.protobuf.ByteString getInputEventProtoBytes(); // required string vertex_name = 2; /** * <code>required string vertex_name = 2;</code> */ boolean hasVertexName(); /** * <code>required string vertex_name = 2;</code> */ java.lang.String getVertexName(); /** * <code>required string vertex_name = 2;</code> */ com.google.protobuf.ByteString getVertexNameBytes(); // required string dest_input_name = 3; /** * <code>required string dest_input_name = 3;</code> */ boolean hasDestInputName(); /** * <code>required string dest_input_name = 3;</code> */ java.lang.String getDestInputName(); /** * <code>required string dest_input_name = 3;</code> */ com.google.protobuf.ByteString getDestInputNameBytes(); // optional int32 key_id = 4; /** * <code>optional int32 key_id = 4;</code> */ boolean hasKeyId(); /** * <code>optional int32 key_id = 4;</code> */ int getKeyId(); } /** * Protobuf type {@code NotTezEvent} * * <pre> ** * Tez API implementation derives an enum value from instanceof on the event, then uses that enum * in a giant switch statement to re-implement virtual functions. In addition the event classes * are final and serialization is a mix of writables and protobuf. As a result, it is impossible * three times over to add anything there. So, we'd do our own "inspired" serialization. * Eventually we'll move away from events for API. * </pre> */ public static final class NotTezEvent extends com.google.protobuf.GeneratedMessage implements NotTezEventOrBuilder { // Use NotTezEvent.newBuilder() to construct. private NotTezEvent(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private NotTezEvent(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final NotTezEvent defaultInstance; public static NotTezEvent getDefaultInstance() { return defaultInstance; } public NotTezEvent getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private NotTezEvent( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { bitField0_ |= 0x00000001; inputEventProtoBytes_ = input.readBytes(); break; } case 18: { bitField0_ |= 0x00000002; vertexName_ = input.readBytes(); break; } case 26: { bitField0_ |= 0x00000004; destInputName_ = input.readBytes(); break; } case 32: { bitField0_ |= 0x00000008; keyId_ = input.readInt32(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_NotTezEvent_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_NotTezEvent_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent.Builder.class); } public static com.google.protobuf.Parser<NotTezEvent> PARSER = new com.google.protobuf.AbstractParser<NotTezEvent>() { public NotTezEvent parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new NotTezEvent(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<NotTezEvent> getParserForType() { return PARSER; } private int bitField0_; // required bytes input_event_proto_bytes = 1; public static final int INPUT_EVENT_PROTO_BYTES_FIELD_NUMBER = 1; private com.google.protobuf.ByteString inputEventProtoBytes_; /** * <code>required bytes input_event_proto_bytes = 1;</code> */ public boolean hasInputEventProtoBytes() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required bytes input_event_proto_bytes = 1;</code> */ public com.google.protobuf.ByteString getInputEventProtoBytes() { return inputEventProtoBytes_; } // required string vertex_name = 2; public static final int VERTEX_NAME_FIELD_NUMBER = 2; private java.lang.Object vertexName_; /** * <code>required string vertex_name = 2;</code> */ public boolean hasVertexName() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>required string vertex_name = 2;</code> */ public java.lang.String getVertexName() { java.lang.Object ref = vertexName_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { vertexName_ = s; } return s; } } /** * <code>required string vertex_name = 2;</code> */ public com.google.protobuf.ByteString getVertexNameBytes() { java.lang.Object ref = vertexName_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); vertexName_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } // required string dest_input_name = 3; public static final int DEST_INPUT_NAME_FIELD_NUMBER = 3; private java.lang.Object destInputName_; /** * <code>required string dest_input_name = 3;</code> */ public boolean hasDestInputName() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>required string dest_input_name = 3;</code> */ public java.lang.String getDestInputName() { java.lang.Object ref = destInputName_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { destInputName_ = s; } return s; } } /** * <code>required string dest_input_name = 3;</code> */ public com.google.protobuf.ByteString getDestInputNameBytes() { java.lang.Object ref = destInputName_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); destInputName_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } // optional int32 key_id = 4; public static final int KEY_ID_FIELD_NUMBER = 4; private int keyId_; /** * <code>optional int32 key_id = 4;</code> */ public boolean hasKeyId() { return ((bitField0_ & 0x00000008) == 0x00000008); } /** * <code>optional int32 key_id = 4;</code> */ public int getKeyId() { return keyId_; } private void initFields() { inputEventProtoBytes_ = com.google.protobuf.ByteString.EMPTY; vertexName_ = ""; destInputName_ = ""; keyId_ = 0; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; if (!hasInputEventProtoBytes()) { memoizedIsInitialized = 0; return false; } if (!hasVertexName()) { memoizedIsInitialized = 0; return false; } if (!hasDestInputName()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeBytes(1, inputEventProtoBytes_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeBytes(2, getVertexNameBytes()); } if (((bitField0_ & 0x00000004) == 0x00000004)) { output.writeBytes(3, getDestInputNameBytes()); } if (((bitField0_ & 0x00000008) == 0x00000008)) { output.writeInt32(4, keyId_); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, inputEventProtoBytes_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(2, getVertexNameBytes()); } if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(3, getDestInputNameBytes()); } if (((bitField0_ & 0x00000008) == 0x00000008)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(4, keyId_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent) obj; boolean result = true; result = result && (hasInputEventProtoBytes() == other.hasInputEventProtoBytes()); if (hasInputEventProtoBytes()) { result = result && getInputEventProtoBytes() .equals(other.getInputEventProtoBytes()); } result = result && (hasVertexName() == other.hasVertexName()); if (hasVertexName()) { result = result && getVertexName() .equals(other.getVertexName()); } result = result && (hasDestInputName() == other.hasDestInputName()); if (hasDestInputName()) { result = result && getDestInputName() .equals(other.getDestInputName()); } result = result && (hasKeyId() == other.hasKeyId()); if (hasKeyId()) { result = result && (getKeyId() == other.getKeyId()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasInputEventProtoBytes()) { hash = (37 * hash) + INPUT_EVENT_PROTO_BYTES_FIELD_NUMBER; hash = (53 * hash) + getInputEventProtoBytes().hashCode(); } if (hasVertexName()) { hash = (37 * hash) + VERTEX_NAME_FIELD_NUMBER; hash = (53 * hash) + getVertexName().hashCode(); } if (hasDestInputName()) { hash = (37 * hash) + DEST_INPUT_NAME_FIELD_NUMBER; hash = (53 * hash) + getDestInputName().hashCode(); } if (hasKeyId()) { hash = (37 * hash) + KEY_ID_FIELD_NUMBER; hash = (53 * hash) + getKeyId(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code NotTezEvent} * * <pre> ** * Tez API implementation derives an enum value from instanceof on the event, then uses that enum * in a giant switch statement to re-implement virtual functions. In addition the event classes * are final and serialization is a mix of writables and protobuf. As a result, it is impossible * three times over to add anything there. So, we'd do our own "inspired" serialization. * Eventually we'll move away from events for API. * </pre> */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEventOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_NotTezEvent_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_NotTezEvent_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); inputEventProtoBytes_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); vertexName_ = ""; bitField0_ = (bitField0_ & ~0x00000002); destInputName_ = ""; bitField0_ = (bitField0_ & ~0x00000004); keyId_ = 0; bitField0_ = (bitField0_ & ~0x00000008); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_NotTezEvent_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.inputEventProtoBytes_ = inputEventProtoBytes_; if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } result.vertexName_ = vertexName_; if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000004; } result.destInputName_ = destInputName_; if (((from_bitField0_ & 0x00000008) == 0x00000008)) { to_bitField0_ |= 0x00000008; } result.keyId_ = keyId_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent.getDefaultInstance()) return this; if (other.hasInputEventProtoBytes()) { setInputEventProtoBytes(other.getInputEventProtoBytes()); } if (other.hasVertexName()) { bitField0_ |= 0x00000002; vertexName_ = other.vertexName_; onChanged(); } if (other.hasDestInputName()) { bitField0_ |= 0x00000004; destInputName_ = other.destInputName_; onChanged(); } if (other.hasKeyId()) { setKeyId(other.getKeyId()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { if (!hasInputEventProtoBytes()) { return false; } if (!hasVertexName()) { return false; } if (!hasDestInputName()) { return false; } return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // required bytes input_event_proto_bytes = 1; private com.google.protobuf.ByteString inputEventProtoBytes_ = com.google.protobuf.ByteString.EMPTY; /** * <code>required bytes input_event_proto_bytes = 1;</code> */ public boolean hasInputEventProtoBytes() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required bytes input_event_proto_bytes = 1;</code> */ public com.google.protobuf.ByteString getInputEventProtoBytes() { return inputEventProtoBytes_; } /** * <code>required bytes input_event_proto_bytes = 1;</code> */ public Builder setInputEventProtoBytes(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; inputEventProtoBytes_ = value; onChanged(); return this; } /** * <code>required bytes input_event_proto_bytes = 1;</code> */ public Builder clearInputEventProtoBytes() { bitField0_ = (bitField0_ & ~0x00000001); inputEventProtoBytes_ = getDefaultInstance().getInputEventProtoBytes(); onChanged(); return this; } // required string vertex_name = 2; private java.lang.Object vertexName_ = ""; /** * <code>required string vertex_name = 2;</code> */ public boolean hasVertexName() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>required string vertex_name = 2;</code> */ public java.lang.String getVertexName() { java.lang.Object ref = vertexName_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); vertexName_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>required string vertex_name = 2;</code> */ public com.google.protobuf.ByteString getVertexNameBytes() { java.lang.Object ref = vertexName_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); vertexName_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>required string vertex_name = 2;</code> */ public Builder setVertexName( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; vertexName_ = value; onChanged(); return this; } /** * <code>required string vertex_name = 2;</code> */ public Builder clearVertexName() { bitField0_ = (bitField0_ & ~0x00000002); vertexName_ = getDefaultInstance().getVertexName(); onChanged(); return this; } /** * <code>required string vertex_name = 2;</code> */ public Builder setVertexNameBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; vertexName_ = value; onChanged(); return this; } // required string dest_input_name = 3; private java.lang.Object destInputName_ = ""; /** * <code>required string dest_input_name = 3;</code> */ public boolean hasDestInputName() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>required string dest_input_name = 3;</code> */ public java.lang.String getDestInputName() { java.lang.Object ref = destInputName_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); destInputName_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>required string dest_input_name = 3;</code> */ public com.google.protobuf.ByteString getDestInputNameBytes() { java.lang.Object ref = destInputName_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); destInputName_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>required string dest_input_name = 3;</code> */ public Builder setDestInputName( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000004; destInputName_ = value; onChanged(); return this; } /** * <code>required string dest_input_name = 3;</code> */ public Builder clearDestInputName() { bitField0_ = (bitField0_ & ~0x00000004); destInputName_ = getDefaultInstance().getDestInputName(); onChanged(); return this; } /** * <code>required string dest_input_name = 3;</code> */ public Builder setDestInputNameBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000004; destInputName_ = value; onChanged(); return this; } // optional int32 key_id = 4; private int keyId_ ; /** * <code>optional int32 key_id = 4;</code> */ public boolean hasKeyId() { return ((bitField0_ & 0x00000008) == 0x00000008); } /** * <code>optional int32 key_id = 4;</code> */ public int getKeyId() { return keyId_; } /** * <code>optional int32 key_id = 4;</code> */ public Builder setKeyId(int value) { bitField0_ |= 0x00000008; keyId_ = value; onChanged(); return this; } /** * <code>optional int32 key_id = 4;</code> */ public Builder clearKeyId() { bitField0_ = (bitField0_ & ~0x00000008); keyId_ = 0; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:NotTezEvent) } static { defaultInstance = new NotTezEvent(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:NotTezEvent) } public interface SubmitWorkRequestProtoOrBuilder extends com.google.protobuf.MessageOrBuilder { // optional .VertexOrBinary work_spec = 1; /** * <code>optional .VertexOrBinary work_spec = 1;</code> */ boolean hasWorkSpec(); /** * <code>optional .VertexOrBinary work_spec = 1;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary getWorkSpec(); /** * <code>optional .VertexOrBinary work_spec = 1;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinaryOrBuilder getWorkSpecOrBuilder(); // optional bytes work_spec_signature = 2; /** * <code>optional bytes work_spec_signature = 2;</code> */ boolean hasWorkSpecSignature(); /** * <code>optional bytes work_spec_signature = 2;</code> */ com.google.protobuf.ByteString getWorkSpecSignature(); // optional int32 fragment_number = 3; /** * <code>optional int32 fragment_number = 3;</code> */ boolean hasFragmentNumber(); /** * <code>optional int32 fragment_number = 3;</code> */ int getFragmentNumber(); // optional int32 attempt_number = 4; /** * <code>optional int32 attempt_number = 4;</code> */ boolean hasAttemptNumber(); /** * <code>optional int32 attempt_number = 4;</code> */ int getAttemptNumber(); // optional string container_id_string = 5; /** * <code>optional string container_id_string = 5;</code> */ boolean hasContainerIdString(); /** * <code>optional string container_id_string = 5;</code> */ java.lang.String getContainerIdString(); /** * <code>optional string container_id_string = 5;</code> */ com.google.protobuf.ByteString getContainerIdStringBytes(); // optional string am_host = 6; /** * <code>optional string am_host = 6;</code> */ boolean hasAmHost(); /** * <code>optional string am_host = 6;</code> */ java.lang.String getAmHost(); /** * <code>optional string am_host = 6;</code> */ com.google.protobuf.ByteString getAmHostBytes(); // optional int32 am_port = 7; /** * <code>optional int32 am_port = 7;</code> */ boolean hasAmPort(); /** * <code>optional int32 am_port = 7;</code> */ int getAmPort(); // optional bytes credentials_binary = 8; /** * <code>optional bytes credentials_binary = 8;</code> * * <pre> * Credentials are not signed - the client can add e.g. his own HDFS tokens. * </pre> */ boolean hasCredentialsBinary(); /** * <code>optional bytes credentials_binary = 8;</code> * * <pre> * Credentials are not signed - the client can add e.g. his own HDFS tokens. * </pre> */ com.google.protobuf.ByteString getCredentialsBinary(); // optional .FragmentRuntimeInfo fragment_runtime_info = 9; /** * <code>optional .FragmentRuntimeInfo fragment_runtime_info = 9;</code> * * <pre> * Not supported/honored for external clients right now. * </pre> */ boolean hasFragmentRuntimeInfo(); /** * <code>optional .FragmentRuntimeInfo fragment_runtime_info = 9;</code> * * <pre> * Not supported/honored for external clients right now. * </pre> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo getFragmentRuntimeInfo(); /** * <code>optional .FragmentRuntimeInfo fragment_runtime_info = 9;</code> * * <pre> * Not supported/honored for external clients right now. * </pre> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder getFragmentRuntimeInfoOrBuilder(); // optional bytes initial_event_bytes = 10; /** * <code>optional bytes initial_event_bytes = 10;</code> * * <pre> * Serialized (and signed) NotTezEvent; used only for external clients for now. * </pre> */ boolean hasInitialEventBytes(); /** * <code>optional bytes initial_event_bytes = 10;</code> * * <pre> * Serialized (and signed) NotTezEvent; used only for external clients for now. * </pre> */ com.google.protobuf.ByteString getInitialEventBytes(); // optional bytes initial_event_signature = 11; /** * <code>optional bytes initial_event_signature = 11;</code> */ boolean hasInitialEventSignature(); /** * <code>optional bytes initial_event_signature = 11;</code> */ com.google.protobuf.ByteString getInitialEventSignature(); } /** * Protobuf type {@code SubmitWorkRequestProto} */ public static final class SubmitWorkRequestProto extends com.google.protobuf.GeneratedMessage implements SubmitWorkRequestProtoOrBuilder { // Use SubmitWorkRequestProto.newBuilder() to construct. private SubmitWorkRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private SubmitWorkRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final SubmitWorkRequestProto defaultInstance; public static SubmitWorkRequestProto getDefaultInstance() { return defaultInstance; } public SubmitWorkRequestProto getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SubmitWorkRequestProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = workSpec_.toBuilder(); } workSpec_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(workSpec_); workSpec_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 18: { bitField0_ |= 0x00000002; workSpecSignature_ = input.readBytes(); break; } case 24: { bitField0_ |= 0x00000004; fragmentNumber_ = input.readInt32(); break; } case 32: { bitField0_ |= 0x00000008; attemptNumber_ = input.readInt32(); break; } case 42: { bitField0_ |= 0x00000010; containerIdString_ = input.readBytes(); break; } case 50: { bitField0_ |= 0x00000020; amHost_ = input.readBytes(); break; } case 56: { bitField0_ |= 0x00000040; amPort_ = input.readInt32(); break; } case 66: { bitField0_ |= 0x00000080; credentialsBinary_ = input.readBytes(); break; } case 74: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.Builder subBuilder = null; if (((bitField0_ & 0x00000100) == 0x00000100)) { subBuilder = fragmentRuntimeInfo_.toBuilder(); } fragmentRuntimeInfo_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(fragmentRuntimeInfo_); fragmentRuntimeInfo_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000100; break; } case 82: { bitField0_ |= 0x00000200; initialEventBytes_ = input.readBytes(); break; } case 90: { bitField0_ |= 0x00000400; initialEventSignature_ = input.readBytes(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkRequestProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto.Builder.class); } public static com.google.protobuf.Parser<SubmitWorkRequestProto> PARSER = new com.google.protobuf.AbstractParser<SubmitWorkRequestProto>() { public SubmitWorkRequestProto parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new SubmitWorkRequestProto(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<SubmitWorkRequestProto> getParserForType() { return PARSER; } private int bitField0_; // optional .VertexOrBinary work_spec = 1; public static final int WORK_SPEC_FIELD_NUMBER = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary workSpec_; /** * <code>optional .VertexOrBinary work_spec = 1;</code> */ public boolean hasWorkSpec() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional .VertexOrBinary work_spec = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary getWorkSpec() { return workSpec_; } /** * <code>optional .VertexOrBinary work_spec = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinaryOrBuilder getWorkSpecOrBuilder() { return workSpec_; } // optional bytes work_spec_signature = 2; public static final int WORK_SPEC_SIGNATURE_FIELD_NUMBER = 2; private com.google.protobuf.ByteString workSpecSignature_; /** * <code>optional bytes work_spec_signature = 2;</code> */ public boolean hasWorkSpecSignature() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional bytes work_spec_signature = 2;</code> */ public com.google.protobuf.ByteString getWorkSpecSignature() { return workSpecSignature_; } // optional int32 fragment_number = 3; public static final int FRAGMENT_NUMBER_FIELD_NUMBER = 3; private int fragmentNumber_; /** * <code>optional int32 fragment_number = 3;</code> */ public boolean hasFragmentNumber() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional int32 fragment_number = 3;</code> */ public int getFragmentNumber() { return fragmentNumber_; } // optional int32 attempt_number = 4; public static final int ATTEMPT_NUMBER_FIELD_NUMBER = 4; private int attemptNumber_; /** * <code>optional int32 attempt_number = 4;</code> */ public boolean hasAttemptNumber() { return ((bitField0_ & 0x00000008) == 0x00000008); } /** * <code>optional int32 attempt_number = 4;</code> */ public int getAttemptNumber() { return attemptNumber_; } // optional string container_id_string = 5; public static final int CONTAINER_ID_STRING_FIELD_NUMBER = 5; private java.lang.Object containerIdString_; /** * <code>optional string container_id_string = 5;</code> */ public boolean hasContainerIdString() { return ((bitField0_ & 0x00000010) == 0x00000010); } /** * <code>optional string container_id_string = 5;</code> */ public java.lang.String getContainerIdString() { java.lang.Object ref = containerIdString_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { containerIdString_ = s; } return s; } } /** * <code>optional string container_id_string = 5;</code> */ public com.google.protobuf.ByteString getContainerIdStringBytes() { java.lang.Object ref = containerIdString_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); containerIdString_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } // optional string am_host = 6; public static final int AM_HOST_FIELD_NUMBER = 6; private java.lang.Object amHost_; /** * <code>optional string am_host = 6;</code> */ public boolean hasAmHost() { return ((bitField0_ & 0x00000020) == 0x00000020); } /** * <code>optional string am_host = 6;</code> */ public java.lang.String getAmHost() { java.lang.Object ref = amHost_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { amHost_ = s; } return s; } } /** * <code>optional string am_host = 6;</code> */ public com.google.protobuf.ByteString getAmHostBytes() { java.lang.Object ref = amHost_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); amHost_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } // optional int32 am_port = 7; public static final int AM_PORT_FIELD_NUMBER = 7; private int amPort_; /** * <code>optional int32 am_port = 7;</code> */ public boolean hasAmPort() { return ((bitField0_ & 0x00000040) == 0x00000040); } /** * <code>optional int32 am_port = 7;</code> */ public int getAmPort() { return amPort_; } // optional bytes credentials_binary = 8; public static final int CREDENTIALS_BINARY_FIELD_NUMBER = 8; private com.google.protobuf.ByteString credentialsBinary_; /** * <code>optional bytes credentials_binary = 8;</code> * * <pre> * Credentials are not signed - the client can add e.g. his own HDFS tokens. * </pre> */ public boolean hasCredentialsBinary() { return ((bitField0_ & 0x00000080) == 0x00000080); } /** * <code>optional bytes credentials_binary = 8;</code> * * <pre> * Credentials are not signed - the client can add e.g. his own HDFS tokens. * </pre> */ public com.google.protobuf.ByteString getCredentialsBinary() { return credentialsBinary_; } // optional .FragmentRuntimeInfo fragment_runtime_info = 9; public static final int FRAGMENT_RUNTIME_INFO_FIELD_NUMBER = 9; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo fragmentRuntimeInfo_; /** * <code>optional .FragmentRuntimeInfo fragment_runtime_info = 9;</code> * * <pre> * Not supported/honored for external clients right now. * </pre> */ public boolean hasFragmentRuntimeInfo() { return ((bitField0_ & 0x00000100) == 0x00000100); } /** * <code>optional .FragmentRuntimeInfo fragment_runtime_info = 9;</code> * * <pre> * Not supported/honored for external clients right now. * </pre> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo getFragmentRuntimeInfo() { return fragmentRuntimeInfo_; } /** * <code>optional .FragmentRuntimeInfo fragment_runtime_info = 9;</code> * * <pre> * Not supported/honored for external clients right now. * </pre> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder getFragmentRuntimeInfoOrBuilder() { return fragmentRuntimeInfo_; } // optional bytes initial_event_bytes = 10; public static final int INITIAL_EVENT_BYTES_FIELD_NUMBER = 10; private com.google.protobuf.ByteString initialEventBytes_; /** * <code>optional bytes initial_event_bytes = 10;</code> * * <pre> * Serialized (and signed) NotTezEvent; used only for external clients for now. * </pre> */ public boolean hasInitialEventBytes() { return ((bitField0_ & 0x00000200) == 0x00000200); } /** * <code>optional bytes initial_event_bytes = 10;</code> * * <pre> * Serialized (and signed) NotTezEvent; used only for external clients for now. * </pre> */ public com.google.protobuf.ByteString getInitialEventBytes() { return initialEventBytes_; } // optional bytes initial_event_signature = 11; public static final int INITIAL_EVENT_SIGNATURE_FIELD_NUMBER = 11; private com.google.protobuf.ByteString initialEventSignature_; /** * <code>optional bytes initial_event_signature = 11;</code> */ public boolean hasInitialEventSignature() { return ((bitField0_ & 0x00000400) == 0x00000400); } /** * <code>optional bytes initial_event_signature = 11;</code> */ public com.google.protobuf.ByteString getInitialEventSignature() { return initialEventSignature_; } private void initFields() { workSpec_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance(); workSpecSignature_ = com.google.protobuf.ByteString.EMPTY; fragmentNumber_ = 0; attemptNumber_ = 0; containerIdString_ = ""; amHost_ = ""; amPort_ = 0; credentialsBinary_ = com.google.protobuf.ByteString.EMPTY; fragmentRuntimeInfo_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance(); initialEventBytes_ = com.google.protobuf.ByteString.EMPTY; initialEventSignature_ = com.google.protobuf.ByteString.EMPTY; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeMessage(1, workSpec_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeBytes(2, workSpecSignature_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { output.writeInt32(3, fragmentNumber_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { output.writeInt32(4, attemptNumber_); } if (((bitField0_ & 0x00000010) == 0x00000010)) { output.writeBytes(5, getContainerIdStringBytes()); } if (((bitField0_ & 0x00000020) == 0x00000020)) { output.writeBytes(6, getAmHostBytes()); } if (((bitField0_ & 0x00000040) == 0x00000040)) { output.writeInt32(7, amPort_); } if (((bitField0_ & 0x00000080) == 0x00000080)) { output.writeBytes(8, credentialsBinary_); } if (((bitField0_ & 0x00000100) == 0x00000100)) { output.writeMessage(9, fragmentRuntimeInfo_); } if (((bitField0_ & 0x00000200) == 0x00000200)) { output.writeBytes(10, initialEventBytes_); } if (((bitField0_ & 0x00000400) == 0x00000400)) { output.writeBytes(11, initialEventSignature_); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(1, workSpec_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(2, workSpecSignature_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(3, fragmentNumber_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(4, attemptNumber_); } if (((bitField0_ & 0x00000010) == 0x00000010)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(5, getContainerIdStringBytes()); } if (((bitField0_ & 0x00000020) == 0x00000020)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(6, getAmHostBytes()); } if (((bitField0_ & 0x00000040) == 0x00000040)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(7, amPort_); } if (((bitField0_ & 0x00000080) == 0x00000080)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(8, credentialsBinary_); } if (((bitField0_ & 0x00000100) == 0x00000100)) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(9, fragmentRuntimeInfo_); } if (((bitField0_ & 0x00000200) == 0x00000200)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(10, initialEventBytes_); } if (((bitField0_ & 0x00000400) == 0x00000400)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(11, initialEventSignature_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto) obj; boolean result = true; result = result && (hasWorkSpec() == other.hasWorkSpec()); if (hasWorkSpec()) { result = result && getWorkSpec() .equals(other.getWorkSpec()); } result = result && (hasWorkSpecSignature() == other.hasWorkSpecSignature()); if (hasWorkSpecSignature()) { result = result && getWorkSpecSignature() .equals(other.getWorkSpecSignature()); } result = result && (hasFragmentNumber() == other.hasFragmentNumber()); if (hasFragmentNumber()) { result = result && (getFragmentNumber() == other.getFragmentNumber()); } result = result && (hasAttemptNumber() == other.hasAttemptNumber()); if (hasAttemptNumber()) { result = result && (getAttemptNumber() == other.getAttemptNumber()); } result = result && (hasContainerIdString() == other.hasContainerIdString()); if (hasContainerIdString()) { result = result && getContainerIdString() .equals(other.getContainerIdString()); } result = result && (hasAmHost() == other.hasAmHost()); if (hasAmHost()) { result = result && getAmHost() .equals(other.getAmHost()); } result = result && (hasAmPort() == other.hasAmPort()); if (hasAmPort()) { result = result && (getAmPort() == other.getAmPort()); } result = result && (hasCredentialsBinary() == other.hasCredentialsBinary()); if (hasCredentialsBinary()) { result = result && getCredentialsBinary() .equals(other.getCredentialsBinary()); } result = result && (hasFragmentRuntimeInfo() == other.hasFragmentRuntimeInfo()); if (hasFragmentRuntimeInfo()) { result = result && getFragmentRuntimeInfo() .equals(other.getFragmentRuntimeInfo()); } result = result && (hasInitialEventBytes() == other.hasInitialEventBytes()); if (hasInitialEventBytes()) { result = result && getInitialEventBytes() .equals(other.getInitialEventBytes()); } result = result && (hasInitialEventSignature() == other.hasInitialEventSignature()); if (hasInitialEventSignature()) { result = result && getInitialEventSignature() .equals(other.getInitialEventSignature()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasWorkSpec()) { hash = (37 * hash) + WORK_SPEC_FIELD_NUMBER; hash = (53 * hash) + getWorkSpec().hashCode(); } if (hasWorkSpecSignature()) { hash = (37 * hash) + WORK_SPEC_SIGNATURE_FIELD_NUMBER; hash = (53 * hash) + getWorkSpecSignature().hashCode(); } if (hasFragmentNumber()) { hash = (37 * hash) + FRAGMENT_NUMBER_FIELD_NUMBER; hash = (53 * hash) + getFragmentNumber(); } if (hasAttemptNumber()) { hash = (37 * hash) + ATTEMPT_NUMBER_FIELD_NUMBER; hash = (53 * hash) + getAttemptNumber(); } if (hasContainerIdString()) { hash = (37 * hash) + CONTAINER_ID_STRING_FIELD_NUMBER; hash = (53 * hash) + getContainerIdString().hashCode(); } if (hasAmHost()) { hash = (37 * hash) + AM_HOST_FIELD_NUMBER; hash = (53 * hash) + getAmHost().hashCode(); } if (hasAmPort()) { hash = (37 * hash) + AM_PORT_FIELD_NUMBER; hash = (53 * hash) + getAmPort(); } if (hasCredentialsBinary()) { hash = (37 * hash) + CREDENTIALS_BINARY_FIELD_NUMBER; hash = (53 * hash) + getCredentialsBinary().hashCode(); } if (hasFragmentRuntimeInfo()) { hash = (37 * hash) + FRAGMENT_RUNTIME_INFO_FIELD_NUMBER; hash = (53 * hash) + getFragmentRuntimeInfo().hashCode(); } if (hasInitialEventBytes()) { hash = (37 * hash) + INITIAL_EVENT_BYTES_FIELD_NUMBER; hash = (53 * hash) + getInitialEventBytes().hashCode(); } if (hasInitialEventSignature()) { hash = (37 * hash) + INITIAL_EVENT_SIGNATURE_FIELD_NUMBER; hash = (53 * hash) + getInitialEventSignature().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code SubmitWorkRequestProto} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkRequestProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getWorkSpecFieldBuilder(); getFragmentRuntimeInfoFieldBuilder(); } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); if (workSpecBuilder_ == null) { workSpec_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance(); } else { workSpecBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); workSpecSignature_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); fragmentNumber_ = 0; bitField0_ = (bitField0_ & ~0x00000004); attemptNumber_ = 0; bitField0_ = (bitField0_ & ~0x00000008); containerIdString_ = ""; bitField0_ = (bitField0_ & ~0x00000010); amHost_ = ""; bitField0_ = (bitField0_ & ~0x00000020); amPort_ = 0; bitField0_ = (bitField0_ & ~0x00000040); credentialsBinary_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000080); if (fragmentRuntimeInfoBuilder_ == null) { fragmentRuntimeInfo_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance(); } else { fragmentRuntimeInfoBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000100); initialEventBytes_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000200); initialEventSignature_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000400); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkRequestProto_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } if (workSpecBuilder_ == null) { result.workSpec_ = workSpec_; } else { result.workSpec_ = workSpecBuilder_.build(); } if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } result.workSpecSignature_ = workSpecSignature_; if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000004; } result.fragmentNumber_ = fragmentNumber_; if (((from_bitField0_ & 0x00000008) == 0x00000008)) { to_bitField0_ |= 0x00000008; } result.attemptNumber_ = attemptNumber_; if (((from_bitField0_ & 0x00000010) == 0x00000010)) { to_bitField0_ |= 0x00000010; } result.containerIdString_ = containerIdString_; if (((from_bitField0_ & 0x00000020) == 0x00000020)) { to_bitField0_ |= 0x00000020; } result.amHost_ = amHost_; if (((from_bitField0_ & 0x00000040) == 0x00000040)) { to_bitField0_ |= 0x00000040; } result.amPort_ = amPort_; if (((from_bitField0_ & 0x00000080) == 0x00000080)) { to_bitField0_ |= 0x00000080; } result.credentialsBinary_ = credentialsBinary_; if (((from_bitField0_ & 0x00000100) == 0x00000100)) { to_bitField0_ |= 0x00000100; } if (fragmentRuntimeInfoBuilder_ == null) { result.fragmentRuntimeInfo_ = fragmentRuntimeInfo_; } else { result.fragmentRuntimeInfo_ = fragmentRuntimeInfoBuilder_.build(); } if (((from_bitField0_ & 0x00000200) == 0x00000200)) { to_bitField0_ |= 0x00000200; } result.initialEventBytes_ = initialEventBytes_; if (((from_bitField0_ & 0x00000400) == 0x00000400)) { to_bitField0_ |= 0x00000400; } result.initialEventSignature_ = initialEventSignature_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto.getDefaultInstance()) return this; if (other.hasWorkSpec()) { mergeWorkSpec(other.getWorkSpec()); } if (other.hasWorkSpecSignature()) { setWorkSpecSignature(other.getWorkSpecSignature()); } if (other.hasFragmentNumber()) { setFragmentNumber(other.getFragmentNumber()); } if (other.hasAttemptNumber()) { setAttemptNumber(other.getAttemptNumber()); } if (other.hasContainerIdString()) { bitField0_ |= 0x00000010; containerIdString_ = other.containerIdString_; onChanged(); } if (other.hasAmHost()) { bitField0_ |= 0x00000020; amHost_ = other.amHost_; onChanged(); } if (other.hasAmPort()) { setAmPort(other.getAmPort()); } if (other.hasCredentialsBinary()) { setCredentialsBinary(other.getCredentialsBinary()); } if (other.hasFragmentRuntimeInfo()) { mergeFragmentRuntimeInfo(other.getFragmentRuntimeInfo()); } if (other.hasInitialEventBytes()) { setInitialEventBytes(other.getInitialEventBytes()); } if (other.hasInitialEventSignature()) { setInitialEventSignature(other.getInitialEventSignature()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // optional .VertexOrBinary work_spec = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary workSpec_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinaryOrBuilder> workSpecBuilder_; /** * <code>optional .VertexOrBinary work_spec = 1;</code> */ public boolean hasWorkSpec() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional .VertexOrBinary work_spec = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary getWorkSpec() { if (workSpecBuilder_ == null) { return workSpec_; } else { return workSpecBuilder_.getMessage(); } } /** * <code>optional .VertexOrBinary work_spec = 1;</code> */ public Builder setWorkSpec(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary value) { if (workSpecBuilder_ == null) { if (value == null) { throw new NullPointerException(); } workSpec_ = value; onChanged(); } else { workSpecBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * <code>optional .VertexOrBinary work_spec = 1;</code> */ public Builder setWorkSpec( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.Builder builderForValue) { if (workSpecBuilder_ == null) { workSpec_ = builderForValue.build(); onChanged(); } else { workSpecBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * <code>optional .VertexOrBinary work_spec = 1;</code> */ public Builder mergeWorkSpec(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary value) { if (workSpecBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && workSpec_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance()) { workSpec_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.newBuilder(workSpec_).mergeFrom(value).buildPartial(); } else { workSpec_ = value; } onChanged(); } else { workSpecBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * <code>optional .VertexOrBinary work_spec = 1;</code> */ public Builder clearWorkSpec() { if (workSpecBuilder_ == null) { workSpec_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance(); onChanged(); } else { workSpecBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * <code>optional .VertexOrBinary work_spec = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.Builder getWorkSpecBuilder() { bitField0_ |= 0x00000001; onChanged(); return getWorkSpecFieldBuilder().getBuilder(); } /** * <code>optional .VertexOrBinary work_spec = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinaryOrBuilder getWorkSpecOrBuilder() { if (workSpecBuilder_ != null) { return workSpecBuilder_.getMessageOrBuilder(); } else { return workSpec_; } } /** * <code>optional .VertexOrBinary work_spec = 1;</code> */ private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinaryOrBuilder> getWorkSpecFieldBuilder() { if (workSpecBuilder_ == null) { workSpecBuilder_ = new com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinaryOrBuilder>( workSpec_, getParentForChildren(), isClean()); workSpec_ = null; } return workSpecBuilder_; } // optional bytes work_spec_signature = 2; private com.google.protobuf.ByteString workSpecSignature_ = com.google.protobuf.ByteString.EMPTY; /** * <code>optional bytes work_spec_signature = 2;</code> */ public boolean hasWorkSpecSignature() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional bytes work_spec_signature = 2;</code> */ public com.google.protobuf.ByteString getWorkSpecSignature() { return workSpecSignature_; } /** * <code>optional bytes work_spec_signature = 2;</code> */ public Builder setWorkSpecSignature(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; workSpecSignature_ = value; onChanged(); return this; } /** * <code>optional bytes work_spec_signature = 2;</code> */ public Builder clearWorkSpecSignature() { bitField0_ = (bitField0_ & ~0x00000002); workSpecSignature_ = getDefaultInstance().getWorkSpecSignature(); onChanged(); return this; } // optional int32 fragment_number = 3; private int fragmentNumber_ ; /** * <code>optional int32 fragment_number = 3;</code> */ public boolean hasFragmentNumber() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional int32 fragment_number = 3;</code> */ public int getFragmentNumber() { return fragmentNumber_; } /** * <code>optional int32 fragment_number = 3;</code> */ public Builder setFragmentNumber(int value) { bitField0_ |= 0x00000004; fragmentNumber_ = value; onChanged(); return this; } /** * <code>optional int32 fragment_number = 3;</code> */ public Builder clearFragmentNumber() { bitField0_ = (bitField0_ & ~0x00000004); fragmentNumber_ = 0; onChanged(); return this; } // optional int32 attempt_number = 4; private int attemptNumber_ ; /** * <code>optional int32 attempt_number = 4;</code> */ public boolean hasAttemptNumber() { return ((bitField0_ & 0x00000008) == 0x00000008); } /** * <code>optional int32 attempt_number = 4;</code> */ public int getAttemptNumber() { return attemptNumber_; } /** * <code>optional int32 attempt_number = 4;</code> */ public Builder setAttemptNumber(int value) { bitField0_ |= 0x00000008; attemptNumber_ = value; onChanged(); return this; } /** * <code>optional int32 attempt_number = 4;</code> */ public Builder clearAttemptNumber() { bitField0_ = (bitField0_ & ~0x00000008); attemptNumber_ = 0; onChanged(); return this; } // optional string container_id_string = 5; private java.lang.Object containerIdString_ = ""; /** * <code>optional string container_id_string = 5;</code> */ public boolean hasContainerIdString() { return ((bitField0_ & 0x00000010) == 0x00000010); } /** * <code>optional string container_id_string = 5;</code> */ public java.lang.String getContainerIdString() { java.lang.Object ref = containerIdString_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); containerIdString_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>optional string container_id_string = 5;</code> */ public com.google.protobuf.ByteString getContainerIdStringBytes() { java.lang.Object ref = containerIdString_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); containerIdString_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>optional string container_id_string = 5;</code> */ public Builder setContainerIdString( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000010; containerIdString_ = value; onChanged(); return this; } /** * <code>optional string container_id_string = 5;</code> */ public Builder clearContainerIdString() { bitField0_ = (bitField0_ & ~0x00000010); containerIdString_ = getDefaultInstance().getContainerIdString(); onChanged(); return this; } /** * <code>optional string container_id_string = 5;</code> */ public Builder setContainerIdStringBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000010; containerIdString_ = value; onChanged(); return this; } // optional string am_host = 6; private java.lang.Object amHost_ = ""; /** * <code>optional string am_host = 6;</code> */ public boolean hasAmHost() { return ((bitField0_ & 0x00000020) == 0x00000020); } /** * <code>optional string am_host = 6;</code> */ public java.lang.String getAmHost() { java.lang.Object ref = amHost_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); amHost_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>optional string am_host = 6;</code> */ public com.google.protobuf.ByteString getAmHostBytes() { java.lang.Object ref = amHost_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); amHost_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>optional string am_host = 6;</code> */ public Builder setAmHost( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000020; amHost_ = value; onChanged(); return this; } /** * <code>optional string am_host = 6;</code> */ public Builder clearAmHost() { bitField0_ = (bitField0_ & ~0x00000020); amHost_ = getDefaultInstance().getAmHost(); onChanged(); return this; } /** * <code>optional string am_host = 6;</code> */ public Builder setAmHostBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000020; amHost_ = value; onChanged(); return this; } // optional int32 am_port = 7; private int amPort_ ; /** * <code>optional int32 am_port = 7;</code> */ public boolean hasAmPort() { return ((bitField0_ & 0x00000040) == 0x00000040); } /** * <code>optional int32 am_port = 7;</code> */ public int getAmPort() { return amPort_; } /** * <code>optional int32 am_port = 7;</code> */ public Builder setAmPort(int value) { bitField0_ |= 0x00000040; amPort_ = value; onChanged(); return this; } /** * <code>optional int32 am_port = 7;</code> */ public Builder clearAmPort() { bitField0_ = (bitField0_ & ~0x00000040); amPort_ = 0; onChanged(); return this; } // optional bytes credentials_binary = 8; private com.google.protobuf.ByteString credentialsBinary_ = com.google.protobuf.ByteString.EMPTY; /** * <code>optional bytes credentials_binary = 8;</code> * * <pre> * Credentials are not signed - the client can add e.g. his own HDFS tokens. * </pre> */ public boolean hasCredentialsBinary() { return ((bitField0_ & 0x00000080) == 0x00000080); } /** * <code>optional bytes credentials_binary = 8;</code> * * <pre> * Credentials are not signed - the client can add e.g. his own HDFS tokens. * </pre> */ public com.google.protobuf.ByteString getCredentialsBinary() { return credentialsBinary_; } /** * <code>optional bytes credentials_binary = 8;</code> * * <pre> * Credentials are not signed - the client can add e.g. his own HDFS tokens. * </pre> */ public Builder setCredentialsBinary(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000080; credentialsBinary_ = value; onChanged(); return this; } /** * <code>optional bytes credentials_binary = 8;</code> * * <pre> * Credentials are not signed - the client can add e.g. his own HDFS tokens. * </pre> */ public Builder clearCredentialsBinary() { bitField0_ = (bitField0_ & ~0x00000080); credentialsBinary_ = getDefaultInstance().getCredentialsBinary(); onChanged(); return this; } // optional .FragmentRuntimeInfo fragment_runtime_info = 9; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo fragmentRuntimeInfo_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder> fragmentRuntimeInfoBuilder_; /** * <code>optional .FragmentRuntimeInfo fragment_runtime_info = 9;</code> * * <pre> * Not supported/honored for external clients right now. * </pre> */ public boolean hasFragmentRuntimeInfo() { return ((bitField0_ & 0x00000100) == 0x00000100); } /** * <code>optional .FragmentRuntimeInfo fragment_runtime_info = 9;</code> * * <pre> * Not supported/honored for external clients right now. * </pre> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo getFragmentRuntimeInfo() { if (fragmentRuntimeInfoBuilder_ == null) { return fragmentRuntimeInfo_; } else { return fragmentRuntimeInfoBuilder_.getMessage(); } } /** * <code>optional .FragmentRuntimeInfo fragment_runtime_info = 9;</code> * * <pre> * Not supported/honored for external clients right now. * </pre> */ public Builder setFragmentRuntimeInfo(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo value) { if (fragmentRuntimeInfoBuilder_ == null) { if (value == null) { throw new NullPointerException(); } fragmentRuntimeInfo_ = value; onChanged(); } else { fragmentRuntimeInfoBuilder_.setMessage(value); } bitField0_ |= 0x00000100; return this; } /** * <code>optional .FragmentRuntimeInfo fragment_runtime_info = 9;</code> * * <pre> * Not supported/honored for external clients right now. * </pre> */ public Builder setFragmentRuntimeInfo( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.Builder builderForValue) { if (fragmentRuntimeInfoBuilder_ == null) { fragmentRuntimeInfo_ = builderForValue.build(); onChanged(); } else { fragmentRuntimeInfoBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000100; return this; } /** * <code>optional .FragmentRuntimeInfo fragment_runtime_info = 9;</code> * * <pre> * Not supported/honored for external clients right now. * </pre> */ public Builder mergeFragmentRuntimeInfo(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo value) { if (fragmentRuntimeInfoBuilder_ == null) { if (((bitField0_ & 0x00000100) == 0x00000100) && fragmentRuntimeInfo_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance()) { fragmentRuntimeInfo_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.newBuilder(fragmentRuntimeInfo_).mergeFrom(value).buildPartial(); } else { fragmentRuntimeInfo_ = value; } onChanged(); } else { fragmentRuntimeInfoBuilder_.mergeFrom(value); } bitField0_ |= 0x00000100; return this; } /** * <code>optional .FragmentRuntimeInfo fragment_runtime_info = 9;</code> * * <pre> * Not supported/honored for external clients right now. * </pre> */ public Builder clearFragmentRuntimeInfo() { if (fragmentRuntimeInfoBuilder_ == null) { fragmentRuntimeInfo_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance(); onChanged(); } else { fragmentRuntimeInfoBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000100); return this; } /** * <code>optional .FragmentRuntimeInfo fragment_runtime_info = 9;</code> * * <pre> * Not supported/honored for external clients right now. * </pre> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.Builder getFragmentRuntimeInfoBuilder() { bitField0_ |= 0x00000100; onChanged(); return getFragmentRuntimeInfoFieldBuilder().getBuilder(); } /** * <code>optional .FragmentRuntimeInfo fragment_runtime_info = 9;</code> * * <pre> * Not supported/honored for external clients right now. * </pre> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder getFragmentRuntimeInfoOrBuilder() { if (fragmentRuntimeInfoBuilder_ != null) { return fragmentRuntimeInfoBuilder_.getMessageOrBuilder(); } else { return fragmentRuntimeInfo_; } } /** * <code>optional .FragmentRuntimeInfo fragment_runtime_info = 9;</code> * * <pre> * Not supported/honored for external clients right now. * </pre> */ private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder> getFragmentRuntimeInfoFieldBuilder() { if (fragmentRuntimeInfoBuilder_ == null) { fragmentRuntimeInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder>( fragmentRuntimeInfo_, getParentForChildren(), isClean()); fragmentRuntimeInfo_ = null; } return fragmentRuntimeInfoBuilder_; } // optional bytes initial_event_bytes = 10; private com.google.protobuf.ByteString initialEventBytes_ = com.google.protobuf.ByteString.EMPTY; /** * <code>optional bytes initial_event_bytes = 10;</code> * * <pre> * Serialized (and signed) NotTezEvent; used only for external clients for now. * </pre> */ public boolean hasInitialEventBytes() { return ((bitField0_ & 0x00000200) == 0x00000200); } /** * <code>optional bytes initial_event_bytes = 10;</code> * * <pre> * Serialized (and signed) NotTezEvent; used only for external clients for now. * </pre> */ public com.google.protobuf.ByteString getInitialEventBytes() { return initialEventBytes_; } /** * <code>optional bytes initial_event_bytes = 10;</code> * * <pre> * Serialized (and signed) NotTezEvent; used only for external clients for now. * </pre> */ public Builder setInitialEventBytes(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000200; initialEventBytes_ = value; onChanged(); return this; } /** * <code>optional bytes initial_event_bytes = 10;</code> * * <pre> * Serialized (and signed) NotTezEvent; used only for external clients for now. * </pre> */ public Builder clearInitialEventBytes() { bitField0_ = (bitField0_ & ~0x00000200); initialEventBytes_ = getDefaultInstance().getInitialEventBytes(); onChanged(); return this; } // optional bytes initial_event_signature = 11; private com.google.protobuf.ByteString initialEventSignature_ = com.google.protobuf.ByteString.EMPTY; /** * <code>optional bytes initial_event_signature = 11;</code> */ public boolean hasInitialEventSignature() { return ((bitField0_ & 0x00000400) == 0x00000400); } /** * <code>optional bytes initial_event_signature = 11;</code> */ public com.google.protobuf.ByteString getInitialEventSignature() { return initialEventSignature_; } /** * <code>optional bytes initial_event_signature = 11;</code> */ public Builder setInitialEventSignature(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000400; initialEventSignature_ = value; onChanged(); return this; } /** * <code>optional bytes initial_event_signature = 11;</code> */ public Builder clearInitialEventSignature() { bitField0_ = (bitField0_ & ~0x00000400); initialEventSignature_ = getDefaultInstance().getInitialEventSignature(); onChanged(); return this; } // @@protoc_insertion_point(builder_scope:SubmitWorkRequestProto) } static { defaultInstance = new SubmitWorkRequestProto(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:SubmitWorkRequestProto) } public interface SubmitWorkResponseProtoOrBuilder extends com.google.protobuf.MessageOrBuilder { // optional .SubmissionStateProto submission_state = 1; /** * <code>optional .SubmissionStateProto submission_state = 1;</code> */ boolean hasSubmissionState(); /** * <code>optional .SubmissionStateProto submission_state = 1;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto getSubmissionState(); // optional string unique_node_id = 2; /** * <code>optional string unique_node_id = 2;</code> */ boolean hasUniqueNodeId(); /** * <code>optional string unique_node_id = 2;</code> */ java.lang.String getUniqueNodeId(); /** * <code>optional string unique_node_id = 2;</code> */ com.google.protobuf.ByteString getUniqueNodeIdBytes(); } /** * Protobuf type {@code SubmitWorkResponseProto} */ public static final class SubmitWorkResponseProto extends com.google.protobuf.GeneratedMessage implements SubmitWorkResponseProtoOrBuilder { // Use SubmitWorkResponseProto.newBuilder() to construct. private SubmitWorkResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private SubmitWorkResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final SubmitWorkResponseProto defaultInstance; public static SubmitWorkResponseProto getDefaultInstance() { return defaultInstance; } public SubmitWorkResponseProto getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SubmitWorkResponseProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 8: { int rawValue = input.readEnum(); org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto value = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(1, rawValue); } else { bitField0_ |= 0x00000001; submissionState_ = value; } break; } case 18: { bitField0_ |= 0x00000002; uniqueNodeId_ = input.readBytes(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkResponseProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto.Builder.class); } public static com.google.protobuf.Parser<SubmitWorkResponseProto> PARSER = new com.google.protobuf.AbstractParser<SubmitWorkResponseProto>() { public SubmitWorkResponseProto parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new SubmitWorkResponseProto(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<SubmitWorkResponseProto> getParserForType() { return PARSER; } private int bitField0_; // optional .SubmissionStateProto submission_state = 1; public static final int SUBMISSION_STATE_FIELD_NUMBER = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto submissionState_; /** * <code>optional .SubmissionStateProto submission_state = 1;</code> */ public boolean hasSubmissionState() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional .SubmissionStateProto submission_state = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto getSubmissionState() { return submissionState_; } // optional string unique_node_id = 2; public static final int UNIQUE_NODE_ID_FIELD_NUMBER = 2; private java.lang.Object uniqueNodeId_; /** * <code>optional string unique_node_id = 2;</code> */ public boolean hasUniqueNodeId() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional string unique_node_id = 2;</code> */ public java.lang.String getUniqueNodeId() { java.lang.Object ref = uniqueNodeId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { uniqueNodeId_ = s; } return s; } } /** * <code>optional string unique_node_id = 2;</code> */ public com.google.protobuf.ByteString getUniqueNodeIdBytes() { java.lang.Object ref = uniqueNodeId_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); uniqueNodeId_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } private void initFields() { submissionState_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto.ACCEPTED; uniqueNodeId_ = ""; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeEnum(1, submissionState_.getNumber()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeBytes(2, getUniqueNodeIdBytes()); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeEnumSize(1, submissionState_.getNumber()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(2, getUniqueNodeIdBytes()); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto) obj; boolean result = true; result = result && (hasSubmissionState() == other.hasSubmissionState()); if (hasSubmissionState()) { result = result && (getSubmissionState() == other.getSubmissionState()); } result = result && (hasUniqueNodeId() == other.hasUniqueNodeId()); if (hasUniqueNodeId()) { result = result && getUniqueNodeId() .equals(other.getUniqueNodeId()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasSubmissionState()) { hash = (37 * hash) + SUBMISSION_STATE_FIELD_NUMBER; hash = (53 * hash) + hashEnum(getSubmissionState()); } if (hasUniqueNodeId()) { hash = (37 * hash) + UNIQUE_NODE_ID_FIELD_NUMBER; hash = (53 * hash) + getUniqueNodeId().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code SubmitWorkResponseProto} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkResponseProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); submissionState_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto.ACCEPTED; bitField0_ = (bitField0_ & ~0x00000001); uniqueNodeId_ = ""; bitField0_ = (bitField0_ & ~0x00000002); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkResponseProto_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.submissionState_ = submissionState_; if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } result.uniqueNodeId_ = uniqueNodeId_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto.getDefaultInstance()) return this; if (other.hasSubmissionState()) { setSubmissionState(other.getSubmissionState()); } if (other.hasUniqueNodeId()) { bitField0_ |= 0x00000002; uniqueNodeId_ = other.uniqueNodeId_; onChanged(); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // optional .SubmissionStateProto submission_state = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto submissionState_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto.ACCEPTED; /** * <code>optional .SubmissionStateProto submission_state = 1;</code> */ public boolean hasSubmissionState() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional .SubmissionStateProto submission_state = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto getSubmissionState() { return submissionState_; } /** * <code>optional .SubmissionStateProto submission_state = 1;</code> */ public Builder setSubmissionState(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; submissionState_ = value; onChanged(); return this; } /** * <code>optional .SubmissionStateProto submission_state = 1;</code> */ public Builder clearSubmissionState() { bitField0_ = (bitField0_ & ~0x00000001); submissionState_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto.ACCEPTED; onChanged(); return this; } // optional string unique_node_id = 2; private java.lang.Object uniqueNodeId_ = ""; /** * <code>optional string unique_node_id = 2;</code> */ public boolean hasUniqueNodeId() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional string unique_node_id = 2;</code> */ public java.lang.String getUniqueNodeId() { java.lang.Object ref = uniqueNodeId_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); uniqueNodeId_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>optional string unique_node_id = 2;</code> */ public com.google.protobuf.ByteString getUniqueNodeIdBytes() { java.lang.Object ref = uniqueNodeId_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); uniqueNodeId_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>optional string unique_node_id = 2;</code> */ public Builder setUniqueNodeId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; uniqueNodeId_ = value; onChanged(); return this; } /** * <code>optional string unique_node_id = 2;</code> */ public Builder clearUniqueNodeId() { bitField0_ = (bitField0_ & ~0x00000002); uniqueNodeId_ = getDefaultInstance().getUniqueNodeId(); onChanged(); return this; } /** * <code>optional string unique_node_id = 2;</code> */ public Builder setUniqueNodeIdBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; uniqueNodeId_ = value; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:SubmitWorkResponseProto) } static { defaultInstance = new SubmitWorkResponseProto(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:SubmitWorkResponseProto) } public interface SourceStateUpdatedRequestProtoOrBuilder extends com.google.protobuf.MessageOrBuilder { // optional .QueryIdentifierProto query_identifier = 1; /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ boolean hasQueryIdentifier(); /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier(); /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder(); // optional string src_name = 2; /** * <code>optional string src_name = 2;</code> */ boolean hasSrcName(); /** * <code>optional string src_name = 2;</code> */ java.lang.String getSrcName(); /** * <code>optional string src_name = 2;</code> */ com.google.protobuf.ByteString getSrcNameBytes(); // optional .SourceStateProto state = 3; /** * <code>optional .SourceStateProto state = 3;</code> */ boolean hasState(); /** * <code>optional .SourceStateProto state = 3;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto getState(); } /** * Protobuf type {@code SourceStateUpdatedRequestProto} */ public static final class SourceStateUpdatedRequestProto extends com.google.protobuf.GeneratedMessage implements SourceStateUpdatedRequestProtoOrBuilder { // Use SourceStateUpdatedRequestProto.newBuilder() to construct. private SourceStateUpdatedRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private SourceStateUpdatedRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final SourceStateUpdatedRequestProto defaultInstance; public static SourceStateUpdatedRequestProto getDefaultInstance() { return defaultInstance; } public SourceStateUpdatedRequestProto getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SourceStateUpdatedRequestProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = queryIdentifier_.toBuilder(); } queryIdentifier_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(queryIdentifier_); queryIdentifier_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 18: { bitField0_ |= 0x00000002; srcName_ = input.readBytes(); break; } case 24: { int rawValue = input.readEnum(); org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto value = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(3, rawValue); } else { bitField0_ |= 0x00000004; state_ = value; } break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedRequestProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto.Builder.class); } public static com.google.protobuf.Parser<SourceStateUpdatedRequestProto> PARSER = new com.google.protobuf.AbstractParser<SourceStateUpdatedRequestProto>() { public SourceStateUpdatedRequestProto parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new SourceStateUpdatedRequestProto(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<SourceStateUpdatedRequestProto> getParserForType() { return PARSER; } private int bitField0_; // optional .QueryIdentifierProto query_identifier = 1; public static final int QUERY_IDENTIFIER_FIELD_NUMBER = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_; /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public boolean hasQueryIdentifier() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { return queryIdentifier_; } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() { return queryIdentifier_; } // optional string src_name = 2; public static final int SRC_NAME_FIELD_NUMBER = 2; private java.lang.Object srcName_; /** * <code>optional string src_name = 2;</code> */ public boolean hasSrcName() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional string src_name = 2;</code> */ public java.lang.String getSrcName() { java.lang.Object ref = srcName_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { srcName_ = s; } return s; } } /** * <code>optional string src_name = 2;</code> */ public com.google.protobuf.ByteString getSrcNameBytes() { java.lang.Object ref = srcName_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); srcName_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } // optional .SourceStateProto state = 3; public static final int STATE_FIELD_NUMBER = 3; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto state_; /** * <code>optional .SourceStateProto state = 3;</code> */ public boolean hasState() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional .SourceStateProto state = 3;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto getState() { return state_; } private void initFields() { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); srcName_ = ""; state_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto.S_SUCCEEDED; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeMessage(1, queryIdentifier_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeBytes(2, getSrcNameBytes()); } if (((bitField0_ & 0x00000004) == 0x00000004)) { output.writeEnum(3, state_.getNumber()); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(1, queryIdentifier_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(2, getSrcNameBytes()); } if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream .computeEnumSize(3, state_.getNumber()); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto) obj; boolean result = true; result = result && (hasQueryIdentifier() == other.hasQueryIdentifier()); if (hasQueryIdentifier()) { result = result && getQueryIdentifier() .equals(other.getQueryIdentifier()); } result = result && (hasSrcName() == other.hasSrcName()); if (hasSrcName()) { result = result && getSrcName() .equals(other.getSrcName()); } result = result && (hasState() == other.hasState()); if (hasState()) { result = result && (getState() == other.getState()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasQueryIdentifier()) { hash = (37 * hash) + QUERY_IDENTIFIER_FIELD_NUMBER; hash = (53 * hash) + getQueryIdentifier().hashCode(); } if (hasSrcName()) { hash = (37 * hash) + SRC_NAME_FIELD_NUMBER; hash = (53 * hash) + getSrcName().hashCode(); } if (hasState()) { hash = (37 * hash) + STATE_FIELD_NUMBER; hash = (53 * hash) + hashEnum(getState()); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code SourceStateUpdatedRequestProto} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedRequestProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getQueryIdentifierFieldBuilder(); } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); if (queryIdentifierBuilder_ == null) { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); } else { queryIdentifierBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); srcName_ = ""; bitField0_ = (bitField0_ & ~0x00000002); state_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto.S_SUCCEEDED; bitField0_ = (bitField0_ & ~0x00000004); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedRequestProto_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } if (queryIdentifierBuilder_ == null) { result.queryIdentifier_ = queryIdentifier_; } else { result.queryIdentifier_ = queryIdentifierBuilder_.build(); } if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } result.srcName_ = srcName_; if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000004; } result.state_ = state_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto.getDefaultInstance()) return this; if (other.hasQueryIdentifier()) { mergeQueryIdentifier(other.getQueryIdentifier()); } if (other.hasSrcName()) { bitField0_ |= 0x00000002; srcName_ = other.srcName_; onChanged(); } if (other.hasState()) { setState(other.getState()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // optional .QueryIdentifierProto query_identifier = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> queryIdentifierBuilder_; /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public boolean hasQueryIdentifier() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { if (queryIdentifierBuilder_ == null) { return queryIdentifier_; } else { return queryIdentifierBuilder_.getMessage(); } } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public Builder setQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto value) { if (queryIdentifierBuilder_ == null) { if (value == null) { throw new NullPointerException(); } queryIdentifier_ = value; onChanged(); } else { queryIdentifierBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public Builder setQueryIdentifier( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder builderForValue) { if (queryIdentifierBuilder_ == null) { queryIdentifier_ = builderForValue.build(); onChanged(); } else { queryIdentifierBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto value) { if (queryIdentifierBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && queryIdentifier_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance()) { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.newBuilder(queryIdentifier_).mergeFrom(value).buildPartial(); } else { queryIdentifier_ = value; } onChanged(); } else { queryIdentifierBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public Builder clearQueryIdentifier() { if (queryIdentifierBuilder_ == null) { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); onChanged(); } else { queryIdentifierBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder getQueryIdentifierBuilder() { bitField0_ |= 0x00000001; onChanged(); return getQueryIdentifierFieldBuilder().getBuilder(); } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() { if (queryIdentifierBuilder_ != null) { return queryIdentifierBuilder_.getMessageOrBuilder(); } else { return queryIdentifier_; } } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> getQueryIdentifierFieldBuilder() { if (queryIdentifierBuilder_ == null) { queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder>( queryIdentifier_, getParentForChildren(), isClean()); queryIdentifier_ = null; } return queryIdentifierBuilder_; } // optional string src_name = 2; private java.lang.Object srcName_ = ""; /** * <code>optional string src_name = 2;</code> */ public boolean hasSrcName() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional string src_name = 2;</code> */ public java.lang.String getSrcName() { java.lang.Object ref = srcName_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); srcName_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>optional string src_name = 2;</code> */ public com.google.protobuf.ByteString getSrcNameBytes() { java.lang.Object ref = srcName_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); srcName_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>optional string src_name = 2;</code> */ public Builder setSrcName( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; srcName_ = value; onChanged(); return this; } /** * <code>optional string src_name = 2;</code> */ public Builder clearSrcName() { bitField0_ = (bitField0_ & ~0x00000002); srcName_ = getDefaultInstance().getSrcName(); onChanged(); return this; } /** * <code>optional string src_name = 2;</code> */ public Builder setSrcNameBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; srcName_ = value; onChanged(); return this; } // optional .SourceStateProto state = 3; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto state_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto.S_SUCCEEDED; /** * <code>optional .SourceStateProto state = 3;</code> */ public boolean hasState() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional .SourceStateProto state = 3;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto getState() { return state_; } /** * <code>optional .SourceStateProto state = 3;</code> */ public Builder setState(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000004; state_ = value; onChanged(); return this; } /** * <code>optional .SourceStateProto state = 3;</code> */ public Builder clearState() { bitField0_ = (bitField0_ & ~0x00000004); state_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto.S_SUCCEEDED; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:SourceStateUpdatedRequestProto) } static { defaultInstance = new SourceStateUpdatedRequestProto(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:SourceStateUpdatedRequestProto) } public interface SourceStateUpdatedResponseProtoOrBuilder extends com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code SourceStateUpdatedResponseProto} */ public static final class SourceStateUpdatedResponseProto extends com.google.protobuf.GeneratedMessage implements SourceStateUpdatedResponseProtoOrBuilder { // Use SourceStateUpdatedResponseProto.newBuilder() to construct. private SourceStateUpdatedResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private SourceStateUpdatedResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final SourceStateUpdatedResponseProto defaultInstance; public static SourceStateUpdatedResponseProto getDefaultInstance() { return defaultInstance; } public SourceStateUpdatedResponseProto getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SourceStateUpdatedResponseProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedResponseProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto.Builder.class); } public static com.google.protobuf.Parser<SourceStateUpdatedResponseProto> PARSER = new com.google.protobuf.AbstractParser<SourceStateUpdatedResponseProto>() { public SourceStateUpdatedResponseProto parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new SourceStateUpdatedResponseProto(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<SourceStateUpdatedResponseProto> getParserForType() { return PARSER; } private void initFields() { } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto) obj; boolean result = true; result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code SourceStateUpdatedResponseProto} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedResponseProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedResponseProto_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto(this); onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto.getDefaultInstance()) return this; this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } // @@protoc_insertion_point(builder_scope:SourceStateUpdatedResponseProto) } static { defaultInstance = new SourceStateUpdatedResponseProto(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:SourceStateUpdatedResponseProto) } public interface QueryCompleteRequestProtoOrBuilder extends com.google.protobuf.MessageOrBuilder { // optional .QueryIdentifierProto query_identifier = 1; /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ boolean hasQueryIdentifier(); /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier(); /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder(); // optional int64 delete_delay = 2 [default = 0]; /** * <code>optional int64 delete_delay = 2 [default = 0];</code> */ boolean hasDeleteDelay(); /** * <code>optional int64 delete_delay = 2 [default = 0];</code> */ long getDeleteDelay(); } /** * Protobuf type {@code QueryCompleteRequestProto} */ public static final class QueryCompleteRequestProto extends com.google.protobuf.GeneratedMessage implements QueryCompleteRequestProtoOrBuilder { // Use QueryCompleteRequestProto.newBuilder() to construct. private QueryCompleteRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private QueryCompleteRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final QueryCompleteRequestProto defaultInstance; public static QueryCompleteRequestProto getDefaultInstance() { return defaultInstance; } public QueryCompleteRequestProto getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private QueryCompleteRequestProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = queryIdentifier_.toBuilder(); } queryIdentifier_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(queryIdentifier_); queryIdentifier_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 16: { bitField0_ |= 0x00000002; deleteDelay_ = input.readInt64(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteRequestProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.Builder.class); } public static com.google.protobuf.Parser<QueryCompleteRequestProto> PARSER = new com.google.protobuf.AbstractParser<QueryCompleteRequestProto>() { public QueryCompleteRequestProto parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new QueryCompleteRequestProto(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<QueryCompleteRequestProto> getParserForType() { return PARSER; } private int bitField0_; // optional .QueryIdentifierProto query_identifier = 1; public static final int QUERY_IDENTIFIER_FIELD_NUMBER = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_; /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public boolean hasQueryIdentifier() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { return queryIdentifier_; } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() { return queryIdentifier_; } // optional int64 delete_delay = 2 [default = 0]; public static final int DELETE_DELAY_FIELD_NUMBER = 2; private long deleteDelay_; /** * <code>optional int64 delete_delay = 2 [default = 0];</code> */ public boolean hasDeleteDelay() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional int64 delete_delay = 2 [default = 0];</code> */ public long getDeleteDelay() { return deleteDelay_; } private void initFields() { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); deleteDelay_ = 0L; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeMessage(1, queryIdentifier_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeInt64(2, deleteDelay_); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(1, queryIdentifier_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(2, deleteDelay_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto) obj; boolean result = true; result = result && (hasQueryIdentifier() == other.hasQueryIdentifier()); if (hasQueryIdentifier()) { result = result && getQueryIdentifier() .equals(other.getQueryIdentifier()); } result = result && (hasDeleteDelay() == other.hasDeleteDelay()); if (hasDeleteDelay()) { result = result && (getDeleteDelay() == other.getDeleteDelay()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasQueryIdentifier()) { hash = (37 * hash) + QUERY_IDENTIFIER_FIELD_NUMBER; hash = (53 * hash) + getQueryIdentifier().hashCode(); } if (hasDeleteDelay()) { hash = (37 * hash) + DELETE_DELAY_FIELD_NUMBER; hash = (53 * hash) + hashLong(getDeleteDelay()); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code QueryCompleteRequestProto} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteRequestProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getQueryIdentifierFieldBuilder(); } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); if (queryIdentifierBuilder_ == null) { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); } else { queryIdentifierBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); deleteDelay_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteRequestProto_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } if (queryIdentifierBuilder_ == null) { result.queryIdentifier_ = queryIdentifier_; } else { result.queryIdentifier_ = queryIdentifierBuilder_.build(); } if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } result.deleteDelay_ = deleteDelay_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.getDefaultInstance()) return this; if (other.hasQueryIdentifier()) { mergeQueryIdentifier(other.getQueryIdentifier()); } if (other.hasDeleteDelay()) { setDeleteDelay(other.getDeleteDelay()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // optional .QueryIdentifierProto query_identifier = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> queryIdentifierBuilder_; /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public boolean hasQueryIdentifier() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { if (queryIdentifierBuilder_ == null) { return queryIdentifier_; } else { return queryIdentifierBuilder_.getMessage(); } } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public Builder setQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto value) { if (queryIdentifierBuilder_ == null) { if (value == null) { throw new NullPointerException(); } queryIdentifier_ = value; onChanged(); } else { queryIdentifierBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public Builder setQueryIdentifier( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder builderForValue) { if (queryIdentifierBuilder_ == null) { queryIdentifier_ = builderForValue.build(); onChanged(); } else { queryIdentifierBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto value) { if (queryIdentifierBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && queryIdentifier_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance()) { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.newBuilder(queryIdentifier_).mergeFrom(value).buildPartial(); } else { queryIdentifier_ = value; } onChanged(); } else { queryIdentifierBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public Builder clearQueryIdentifier() { if (queryIdentifierBuilder_ == null) { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); onChanged(); } else { queryIdentifierBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder getQueryIdentifierBuilder() { bitField0_ |= 0x00000001; onChanged(); return getQueryIdentifierFieldBuilder().getBuilder(); } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() { if (queryIdentifierBuilder_ != null) { return queryIdentifierBuilder_.getMessageOrBuilder(); } else { return queryIdentifier_; } } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> getQueryIdentifierFieldBuilder() { if (queryIdentifierBuilder_ == null) { queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder>( queryIdentifier_, getParentForChildren(), isClean()); queryIdentifier_ = null; } return queryIdentifierBuilder_; } // optional int64 delete_delay = 2 [default = 0]; private long deleteDelay_ ; /** * <code>optional int64 delete_delay = 2 [default = 0];</code> */ public boolean hasDeleteDelay() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional int64 delete_delay = 2 [default = 0];</code> */ public long getDeleteDelay() { return deleteDelay_; } /** * <code>optional int64 delete_delay = 2 [default = 0];</code> */ public Builder setDeleteDelay(long value) { bitField0_ |= 0x00000002; deleteDelay_ = value; onChanged(); return this; } /** * <code>optional int64 delete_delay = 2 [default = 0];</code> */ public Builder clearDeleteDelay() { bitField0_ = (bitField0_ & ~0x00000002); deleteDelay_ = 0L; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:QueryCompleteRequestProto) } static { defaultInstance = new QueryCompleteRequestProto(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:QueryCompleteRequestProto) } public interface QueryCompleteResponseProtoOrBuilder extends com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code QueryCompleteResponseProto} */ public static final class QueryCompleteResponseProto extends com.google.protobuf.GeneratedMessage implements QueryCompleteResponseProtoOrBuilder { // Use QueryCompleteResponseProto.newBuilder() to construct. private QueryCompleteResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private QueryCompleteResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final QueryCompleteResponseProto defaultInstance; public static QueryCompleteResponseProto getDefaultInstance() { return defaultInstance; } public QueryCompleteResponseProto getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private QueryCompleteResponseProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteResponseProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.Builder.class); } public static com.google.protobuf.Parser<QueryCompleteResponseProto> PARSER = new com.google.protobuf.AbstractParser<QueryCompleteResponseProto>() { public QueryCompleteResponseProto parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new QueryCompleteResponseProto(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<QueryCompleteResponseProto> getParserForType() { return PARSER; } private void initFields() { } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto) obj; boolean result = true; result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code QueryCompleteResponseProto} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteResponseProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteResponseProto_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto(this); onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.getDefaultInstance()) return this; this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } // @@protoc_insertion_point(builder_scope:QueryCompleteResponseProto) } static { defaultInstance = new QueryCompleteResponseProto(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:QueryCompleteResponseProto) } public interface TerminateFragmentRequestProtoOrBuilder extends com.google.protobuf.MessageOrBuilder { // optional .QueryIdentifierProto query_identifier = 1; /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ boolean hasQueryIdentifier(); /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier(); /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder(); // optional string fragment_identifier_string = 2; /** * <code>optional string fragment_identifier_string = 2;</code> */ boolean hasFragmentIdentifierString(); /** * <code>optional string fragment_identifier_string = 2;</code> */ java.lang.String getFragmentIdentifierString(); /** * <code>optional string fragment_identifier_string = 2;</code> */ com.google.protobuf.ByteString getFragmentIdentifierStringBytes(); } /** * Protobuf type {@code TerminateFragmentRequestProto} */ public static final class TerminateFragmentRequestProto extends com.google.protobuf.GeneratedMessage implements TerminateFragmentRequestProtoOrBuilder { // Use TerminateFragmentRequestProto.newBuilder() to construct. private TerminateFragmentRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private TerminateFragmentRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final TerminateFragmentRequestProto defaultInstance; public static TerminateFragmentRequestProto getDefaultInstance() { return defaultInstance; } public TerminateFragmentRequestProto getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private TerminateFragmentRequestProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = queryIdentifier_.toBuilder(); } queryIdentifier_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(queryIdentifier_); queryIdentifier_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 18: { bitField0_ |= 0x00000002; fragmentIdentifierString_ = input.readBytes(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentRequestProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.Builder.class); } public static com.google.protobuf.Parser<TerminateFragmentRequestProto> PARSER = new com.google.protobuf.AbstractParser<TerminateFragmentRequestProto>() { public TerminateFragmentRequestProto parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new TerminateFragmentRequestProto(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<TerminateFragmentRequestProto> getParserForType() { return PARSER; } private int bitField0_; // optional .QueryIdentifierProto query_identifier = 1; public static final int QUERY_IDENTIFIER_FIELD_NUMBER = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_; /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public boolean hasQueryIdentifier() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { return queryIdentifier_; } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() { return queryIdentifier_; } // optional string fragment_identifier_string = 2; public static final int FRAGMENT_IDENTIFIER_STRING_FIELD_NUMBER = 2; private java.lang.Object fragmentIdentifierString_; /** * <code>optional string fragment_identifier_string = 2;</code> */ public boolean hasFragmentIdentifierString() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional string fragment_identifier_string = 2;</code> */ public java.lang.String getFragmentIdentifierString() { java.lang.Object ref = fragmentIdentifierString_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { fragmentIdentifierString_ = s; } return s; } } /** * <code>optional string fragment_identifier_string = 2;</code> */ public com.google.protobuf.ByteString getFragmentIdentifierStringBytes() { java.lang.Object ref = fragmentIdentifierString_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); fragmentIdentifierString_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } private void initFields() { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); fragmentIdentifierString_ = ""; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeMessage(1, queryIdentifier_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeBytes(2, getFragmentIdentifierStringBytes()); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(1, queryIdentifier_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(2, getFragmentIdentifierStringBytes()); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto) obj; boolean result = true; result = result && (hasQueryIdentifier() == other.hasQueryIdentifier()); if (hasQueryIdentifier()) { result = result && getQueryIdentifier() .equals(other.getQueryIdentifier()); } result = result && (hasFragmentIdentifierString() == other.hasFragmentIdentifierString()); if (hasFragmentIdentifierString()) { result = result && getFragmentIdentifierString() .equals(other.getFragmentIdentifierString()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasQueryIdentifier()) { hash = (37 * hash) + QUERY_IDENTIFIER_FIELD_NUMBER; hash = (53 * hash) + getQueryIdentifier().hashCode(); } if (hasFragmentIdentifierString()) { hash = (37 * hash) + FRAGMENT_IDENTIFIER_STRING_FIELD_NUMBER; hash = (53 * hash) + getFragmentIdentifierString().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code TerminateFragmentRequestProto} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentRequestProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getQueryIdentifierFieldBuilder(); } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); if (queryIdentifierBuilder_ == null) { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); } else { queryIdentifierBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); fragmentIdentifierString_ = ""; bitField0_ = (bitField0_ & ~0x00000002); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentRequestProto_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } if (queryIdentifierBuilder_ == null) { result.queryIdentifier_ = queryIdentifier_; } else { result.queryIdentifier_ = queryIdentifierBuilder_.build(); } if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } result.fragmentIdentifierString_ = fragmentIdentifierString_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.getDefaultInstance()) return this; if (other.hasQueryIdentifier()) { mergeQueryIdentifier(other.getQueryIdentifier()); } if (other.hasFragmentIdentifierString()) { bitField0_ |= 0x00000002; fragmentIdentifierString_ = other.fragmentIdentifierString_; onChanged(); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // optional .QueryIdentifierProto query_identifier = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> queryIdentifierBuilder_; /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public boolean hasQueryIdentifier() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { if (queryIdentifierBuilder_ == null) { return queryIdentifier_; } else { return queryIdentifierBuilder_.getMessage(); } } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public Builder setQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto value) { if (queryIdentifierBuilder_ == null) { if (value == null) { throw new NullPointerException(); } queryIdentifier_ = value; onChanged(); } else { queryIdentifierBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public Builder setQueryIdentifier( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder builderForValue) { if (queryIdentifierBuilder_ == null) { queryIdentifier_ = builderForValue.build(); onChanged(); } else { queryIdentifierBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto value) { if (queryIdentifierBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && queryIdentifier_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance()) { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.newBuilder(queryIdentifier_).mergeFrom(value).buildPartial(); } else { queryIdentifier_ = value; } onChanged(); } else { queryIdentifierBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public Builder clearQueryIdentifier() { if (queryIdentifierBuilder_ == null) { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); onChanged(); } else { queryIdentifierBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder getQueryIdentifierBuilder() { bitField0_ |= 0x00000001; onChanged(); return getQueryIdentifierFieldBuilder().getBuilder(); } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() { if (queryIdentifierBuilder_ != null) { return queryIdentifierBuilder_.getMessageOrBuilder(); } else { return queryIdentifier_; } } /** * <code>optional .QueryIdentifierProto query_identifier = 1;</code> */ private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> getQueryIdentifierFieldBuilder() { if (queryIdentifierBuilder_ == null) { queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder>( queryIdentifier_, getParentForChildren(), isClean()); queryIdentifier_ = null; } return queryIdentifierBuilder_; } // optional string fragment_identifier_string = 2; private java.lang.Object fragmentIdentifierString_ = ""; /** * <code>optional string fragment_identifier_string = 2;</code> */ public boolean hasFragmentIdentifierString() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional string fragment_identifier_string = 2;</code> */ public java.lang.String getFragmentIdentifierString() { java.lang.Object ref = fragmentIdentifierString_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); fragmentIdentifierString_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>optional string fragment_identifier_string = 2;</code> */ public com.google.protobuf.ByteString getFragmentIdentifierStringBytes() { java.lang.Object ref = fragmentIdentifierString_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); fragmentIdentifierString_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>optional string fragment_identifier_string = 2;</code> */ public Builder setFragmentIdentifierString( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; fragmentIdentifierString_ = value; onChanged(); return this; } /** * <code>optional string fragment_identifier_string = 2;</code> */ public Builder clearFragmentIdentifierString() { bitField0_ = (bitField0_ & ~0x00000002); fragmentIdentifierString_ = getDefaultInstance().getFragmentIdentifierString(); onChanged(); return this; } /** * <code>optional string fragment_identifier_string = 2;</code> */ public Builder setFragmentIdentifierStringBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; fragmentIdentifierString_ = value; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:TerminateFragmentRequestProto) } static { defaultInstance = new TerminateFragmentRequestProto(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:TerminateFragmentRequestProto) } public interface TerminateFragmentResponseProtoOrBuilder extends com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code TerminateFragmentResponseProto} */ public static final class TerminateFragmentResponseProto extends com.google.protobuf.GeneratedMessage implements TerminateFragmentResponseProtoOrBuilder { // Use TerminateFragmentResponseProto.newBuilder() to construct. private TerminateFragmentResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private TerminateFragmentResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final TerminateFragmentResponseProto defaultInstance; public static TerminateFragmentResponseProto getDefaultInstance() { return defaultInstance; } public TerminateFragmentResponseProto getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private TerminateFragmentResponseProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentResponseProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.Builder.class); } public static com.google.protobuf.Parser<TerminateFragmentResponseProto> PARSER = new com.google.protobuf.AbstractParser<TerminateFragmentResponseProto>() { public TerminateFragmentResponseProto parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new TerminateFragmentResponseProto(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<TerminateFragmentResponseProto> getParserForType() { return PARSER; } private void initFields() { } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto) obj; boolean result = true; result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code TerminateFragmentResponseProto} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentResponseProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentResponseProto_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto(this); onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.getDefaultInstance()) return this; this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } // @@protoc_insertion_point(builder_scope:TerminateFragmentResponseProto) } static { defaultInstance = new TerminateFragmentResponseProto(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:TerminateFragmentResponseProto) } public interface GetTokenRequestProtoOrBuilder extends com.google.protobuf.MessageOrBuilder { // optional string app_id = 1; /** * <code>optional string app_id = 1;</code> */ boolean hasAppId(); /** * <code>optional string app_id = 1;</code> */ java.lang.String getAppId(); /** * <code>optional string app_id = 1;</code> */ com.google.protobuf.ByteString getAppIdBytes(); } /** * Protobuf type {@code GetTokenRequestProto} */ public static final class GetTokenRequestProto extends com.google.protobuf.GeneratedMessage implements GetTokenRequestProtoOrBuilder { // Use GetTokenRequestProto.newBuilder() to construct. private GetTokenRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private GetTokenRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final GetTokenRequestProto defaultInstance; public static GetTokenRequestProto getDefaultInstance() { return defaultInstance; } public GetTokenRequestProto getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetTokenRequestProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { bitField0_ |= 0x00000001; appId_ = input.readBytes(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenRequestProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto.Builder.class); } public static com.google.protobuf.Parser<GetTokenRequestProto> PARSER = new com.google.protobuf.AbstractParser<GetTokenRequestProto>() { public GetTokenRequestProto parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new GetTokenRequestProto(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<GetTokenRequestProto> getParserForType() { return PARSER; } private int bitField0_; // optional string app_id = 1; public static final int APP_ID_FIELD_NUMBER = 1; private java.lang.Object appId_; /** * <code>optional string app_id = 1;</code> */ public boolean hasAppId() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional string app_id = 1;</code> */ public java.lang.String getAppId() { java.lang.Object ref = appId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { appId_ = s; } return s; } } /** * <code>optional string app_id = 1;</code> */ public com.google.protobuf.ByteString getAppIdBytes() { java.lang.Object ref = appId_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); appId_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } private void initFields() { appId_ = ""; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeBytes(1, getAppIdBytes()); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, getAppIdBytes()); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto) obj; boolean result = true; result = result && (hasAppId() == other.hasAppId()); if (hasAppId()) { result = result && getAppId() .equals(other.getAppId()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasAppId()) { hash = (37 * hash) + APP_ID_FIELD_NUMBER; hash = (53 * hash) + getAppId().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code GetTokenRequestProto} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenRequestProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); appId_ = ""; bitField0_ = (bitField0_ & ~0x00000001); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenRequestProto_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.appId_ = appId_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto.getDefaultInstance()) return this; if (other.hasAppId()) { bitField0_ |= 0x00000001; appId_ = other.appId_; onChanged(); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // optional string app_id = 1; private java.lang.Object appId_ = ""; /** * <code>optional string app_id = 1;</code> */ public boolean hasAppId() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional string app_id = 1;</code> */ public java.lang.String getAppId() { java.lang.Object ref = appId_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); appId_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>optional string app_id = 1;</code> */ public com.google.protobuf.ByteString getAppIdBytes() { java.lang.Object ref = appId_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); appId_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>optional string app_id = 1;</code> */ public Builder setAppId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; appId_ = value; onChanged(); return this; } /** * <code>optional string app_id = 1;</code> */ public Builder clearAppId() { bitField0_ = (bitField0_ & ~0x00000001); appId_ = getDefaultInstance().getAppId(); onChanged(); return this; } /** * <code>optional string app_id = 1;</code> */ public Builder setAppIdBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; appId_ = value; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:GetTokenRequestProto) } static { defaultInstance = new GetTokenRequestProto(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:GetTokenRequestProto) } public interface GetTokenResponseProtoOrBuilder extends com.google.protobuf.MessageOrBuilder { // optional bytes token = 1; /** * <code>optional bytes token = 1;</code> */ boolean hasToken(); /** * <code>optional bytes token = 1;</code> */ com.google.protobuf.ByteString getToken(); } /** * Protobuf type {@code GetTokenResponseProto} */ public static final class GetTokenResponseProto extends com.google.protobuf.GeneratedMessage implements GetTokenResponseProtoOrBuilder { // Use GetTokenResponseProto.newBuilder() to construct. private GetTokenResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private GetTokenResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final GetTokenResponseProto defaultInstance; public static GetTokenResponseProto getDefaultInstance() { return defaultInstance; } public GetTokenResponseProto getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetTokenResponseProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { bitField0_ |= 0x00000001; token_ = input.readBytes(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenResponseProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto.Builder.class); } public static com.google.protobuf.Parser<GetTokenResponseProto> PARSER = new com.google.protobuf.AbstractParser<GetTokenResponseProto>() { public GetTokenResponseProto parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new GetTokenResponseProto(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<GetTokenResponseProto> getParserForType() { return PARSER; } private int bitField0_; // optional bytes token = 1; public static final int TOKEN_FIELD_NUMBER = 1; private com.google.protobuf.ByteString token_; /** * <code>optional bytes token = 1;</code> */ public boolean hasToken() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional bytes token = 1;</code> */ public com.google.protobuf.ByteString getToken() { return token_; } private void initFields() { token_ = com.google.protobuf.ByteString.EMPTY; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeBytes(1, token_); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, token_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto) obj; boolean result = true; result = result && (hasToken() == other.hasToken()); if (hasToken()) { result = result && getToken() .equals(other.getToken()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasToken()) { hash = (37 * hash) + TOKEN_FIELD_NUMBER; hash = (53 * hash) + getToken().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code GetTokenResponseProto} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenResponseProto_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); token_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenResponseProto_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.token_ = token_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto.getDefaultInstance()) return this; if (other.hasToken()) { setToken(other.getToken()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // optional bytes token = 1; private com.google.protobuf.ByteString token_ = com.google.protobuf.ByteString.EMPTY; /** * <code>optional bytes token = 1;</code> */ public boolean hasToken() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional bytes token = 1;</code> */ public com.google.protobuf.ByteString getToken() { return token_; } /** * <code>optional bytes token = 1;</code> */ public Builder setToken(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; token_ = value; onChanged(); return this; } /** * <code>optional bytes token = 1;</code> */ public Builder clearToken() { bitField0_ = (bitField0_ & ~0x00000001); token_ = getDefaultInstance().getToken(); onChanged(); return this; } // @@protoc_insertion_point(builder_scope:GetTokenResponseProto) } static { defaultInstance = new GetTokenResponseProto(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:GetTokenResponseProto) } public interface LlapOutputSocketInitMessageOrBuilder extends com.google.protobuf.MessageOrBuilder { // required string fragment_id = 1; /** * <code>required string fragment_id = 1;</code> */ boolean hasFragmentId(); /** * <code>required string fragment_id = 1;</code> */ java.lang.String getFragmentId(); /** * <code>required string fragment_id = 1;</code> */ com.google.protobuf.ByteString getFragmentIdBytes(); // optional bytes token = 2; /** * <code>optional bytes token = 2;</code> */ boolean hasToken(); /** * <code>optional bytes token = 2;</code> */ com.google.protobuf.ByteString getToken(); } /** * Protobuf type {@code LlapOutputSocketInitMessage} * * <pre> * The message sent by external client to claim the output from the output socket. * </pre> */ public static final class LlapOutputSocketInitMessage extends com.google.protobuf.GeneratedMessage implements LlapOutputSocketInitMessageOrBuilder { // Use LlapOutputSocketInitMessage.newBuilder() to construct. private LlapOutputSocketInitMessage(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private LlapOutputSocketInitMessage(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final LlapOutputSocketInitMessage defaultInstance; public static LlapOutputSocketInitMessage getDefaultInstance() { return defaultInstance; } public LlapOutputSocketInitMessage getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private LlapOutputSocketInitMessage( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { bitField0_ |= 0x00000001; fragmentId_ = input.readBytes(); break; } case 18: { bitField0_ |= 0x00000002; token_ = input.readBytes(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_LlapOutputSocketInitMessage_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_LlapOutputSocketInitMessage_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage.Builder.class); } public static com.google.protobuf.Parser<LlapOutputSocketInitMessage> PARSER = new com.google.protobuf.AbstractParser<LlapOutputSocketInitMessage>() { public LlapOutputSocketInitMessage parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new LlapOutputSocketInitMessage(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<LlapOutputSocketInitMessage> getParserForType() { return PARSER; } private int bitField0_; // required string fragment_id = 1; public static final int FRAGMENT_ID_FIELD_NUMBER = 1; private java.lang.Object fragmentId_; /** * <code>required string fragment_id = 1;</code> */ public boolean hasFragmentId() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required string fragment_id = 1;</code> */ public java.lang.String getFragmentId() { java.lang.Object ref = fragmentId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { fragmentId_ = s; } return s; } } /** * <code>required string fragment_id = 1;</code> */ public com.google.protobuf.ByteString getFragmentIdBytes() { java.lang.Object ref = fragmentId_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); fragmentId_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } // optional bytes token = 2; public static final int TOKEN_FIELD_NUMBER = 2; private com.google.protobuf.ByteString token_; /** * <code>optional bytes token = 2;</code> */ public boolean hasToken() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional bytes token = 2;</code> */ public com.google.protobuf.ByteString getToken() { return token_; } private void initFields() { fragmentId_ = ""; token_ = com.google.protobuf.ByteString.EMPTY; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; if (!hasFragmentId()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeBytes(1, getFragmentIdBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeBytes(2, token_); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, getFragmentIdBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(2, token_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage)) { return super.equals(obj); } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage) obj; boolean result = true; result = result && (hasFragmentId() == other.hasFragmentId()); if (hasFragmentId()) { result = result && getFragmentId() .equals(other.getFragmentId()); } result = result && (hasToken() == other.hasToken()); if (hasToken()) { result = result && getToken() .equals(other.getToken()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasFragmentId()) { hash = (37 * hash) + FRAGMENT_ID_FIELD_NUMBER; hash = (53 * hash) + getFragmentId().hashCode(); } if (hasToken()) { hash = (37 * hash) + TOKEN_FIELD_NUMBER; hash = (53 * hash) + getToken().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code LlapOutputSocketInitMessage} * * <pre> * The message sent by external client to claim the output from the output socket. * </pre> */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessageOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_LlapOutputSocketInitMessage_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_LlapOutputSocketInitMessage_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage.Builder.class); } // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); fragmentId_ = ""; bitField0_ = (bitField0_ & ~0x00000001); token_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_LlapOutputSocketInitMessage_descriptor; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage.getDefaultInstance(); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.fragmentId_ = fragmentId_; if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } result.token_ = token_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage.getDefaultInstance()) return this; if (other.hasFragmentId()) { bitField0_ |= 0x00000001; fragmentId_ = other.fragmentId_; onChanged(); } if (other.hasToken()) { setToken(other.getToken()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { if (!hasFragmentId()) { return false; } return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // required string fragment_id = 1; private java.lang.Object fragmentId_ = ""; /** * <code>required string fragment_id = 1;</code> */ public boolean hasFragmentId() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required string fragment_id = 1;</code> */ public java.lang.String getFragmentId() { java.lang.Object ref = fragmentId_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); fragmentId_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>required string fragment_id = 1;</code> */ public com.google.protobuf.ByteString getFragmentIdBytes() { java.lang.Object ref = fragmentId_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); fragmentId_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>required string fragment_id = 1;</code> */ public Builder setFragmentId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; fragmentId_ = value; onChanged(); return this; } /** * <code>required string fragment_id = 1;</code> */ public Builder clearFragmentId() { bitField0_ = (bitField0_ & ~0x00000001); fragmentId_ = getDefaultInstance().getFragmentId(); onChanged(); return this; } /** * <code>required string fragment_id = 1;</code> */ public Builder setFragmentIdBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; fragmentId_ = value; onChanged(); return this; } // optional bytes token = 2; private com.google.protobuf.ByteString token_ = com.google.protobuf.ByteString.EMPTY; /** * <code>optional bytes token = 2;</code> */ public boolean hasToken() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional bytes token = 2;</code> */ public com.google.protobuf.ByteString getToken() { return token_; } /** * <code>optional bytes token = 2;</code> */ public Builder setToken(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; token_ = value; onChanged(); return this; } /** * <code>optional bytes token = 2;</code> */ public Builder clearToken() { bitField0_ = (bitField0_ & ~0x00000002); token_ = getDefaultInstance().getToken(); onChanged(); return this; } // @@protoc_insertion_point(builder_scope:LlapOutputSocketInitMessage) } static { defaultInstance = new LlapOutputSocketInitMessage(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:LlapOutputSocketInitMessage) } /** * Protobuf service {@code LlapDaemonProtocol} */ public static abstract class LlapDaemonProtocol implements com.google.protobuf.Service { protected LlapDaemonProtocol() {} public interface Interface { /** * <code>rpc submitWork(.SubmitWorkRequestProto) returns (.SubmitWorkResponseProto);</code> */ public abstract void submitWork( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto> done); /** * <code>rpc sourceStateUpdated(.SourceStateUpdatedRequestProto) returns (.SourceStateUpdatedResponseProto);</code> */ public abstract void sourceStateUpdated( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto> done); /** * <code>rpc queryComplete(.QueryCompleteRequestProto) returns (.QueryCompleteResponseProto);</code> */ public abstract void queryComplete( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto> done); /** * <code>rpc terminateFragment(.TerminateFragmentRequestProto) returns (.TerminateFragmentResponseProto);</code> */ public abstract void terminateFragment( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto> done); } public static com.google.protobuf.Service newReflectiveService( final Interface impl) { return new LlapDaemonProtocol() { @java.lang.Override public void submitWork( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto> done) { impl.submitWork(controller, request, done); } @java.lang.Override public void sourceStateUpdated( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto> done) { impl.sourceStateUpdated(controller, request, done); } @java.lang.Override public void queryComplete( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto> done) { impl.queryComplete(controller, request, done); } @java.lang.Override public void terminateFragment( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto> done) { impl.terminateFragment(controller, request, done); } }; } public static com.google.protobuf.BlockingService newReflectiveBlockingService(final BlockingInterface impl) { return new com.google.protobuf.BlockingService() { public final com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final com.google.protobuf.Message callBlockingMethod( com.google.protobuf.Descriptors.MethodDescriptor method, com.google.protobuf.RpcController controller, com.google.protobuf.Message request) throws com.google.protobuf.ServiceException { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callBlockingMethod() given method descriptor for " + "wrong service type."); } switch(method.getIndex()) { case 0: return impl.submitWork(controller, (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto)request); case 1: return impl.sourceStateUpdated(controller, (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto)request); case 2: return impl.queryComplete(controller, (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto)request); case 3: return impl.terminateFragment(controller, (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto)request); default: throw new java.lang.AssertionError("Can't get here."); } } public final com.google.protobuf.Message getRequestPrototype( com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto.getDefaultInstance(); case 1: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto.getDefaultInstance(); case 2: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.getDefaultInstance(); case 3: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } public final com.google.protobuf.Message getResponsePrototype( com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto.getDefaultInstance(); case 1: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto.getDefaultInstance(); case 2: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.getDefaultInstance(); case 3: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } }; } /** * <code>rpc submitWork(.SubmitWorkRequestProto) returns (.SubmitWorkResponseProto);</code> */ public abstract void submitWork( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto> done); /** * <code>rpc sourceStateUpdated(.SourceStateUpdatedRequestProto) returns (.SourceStateUpdatedResponseProto);</code> */ public abstract void sourceStateUpdated( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto> done); /** * <code>rpc queryComplete(.QueryCompleteRequestProto) returns (.QueryCompleteResponseProto);</code> */ public abstract void queryComplete( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto> done); /** * <code>rpc terminateFragment(.TerminateFragmentRequestProto) returns (.TerminateFragmentResponseProto);</code> */ public abstract void terminateFragment( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto> done); public static final com.google.protobuf.Descriptors.ServiceDescriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.getDescriptor().getServices().get(0); } public final com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final void callMethod( com.google.protobuf.Descriptors.MethodDescriptor method, com.google.protobuf.RpcController controller, com.google.protobuf.Message request, com.google.protobuf.RpcCallback< com.google.protobuf.Message> done) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callMethod() given method descriptor for wrong " + "service type."); } switch(method.getIndex()) { case 0: this.submitWork(controller, (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto)request, com.google.protobuf.RpcUtil.<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto>specializeCallback( done)); return; case 1: this.sourceStateUpdated(controller, (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto)request, com.google.protobuf.RpcUtil.<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto>specializeCallback( done)); return; case 2: this.queryComplete(controller, (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto)request, com.google.protobuf.RpcUtil.<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto>specializeCallback( done)); return; case 3: this.terminateFragment(controller, (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto)request, com.google.protobuf.RpcUtil.<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto>specializeCallback( done)); return; default: throw new java.lang.AssertionError("Can't get here."); } } public final com.google.protobuf.Message getRequestPrototype( com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto.getDefaultInstance(); case 1: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto.getDefaultInstance(); case 2: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.getDefaultInstance(); case 3: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } public final com.google.protobuf.Message getResponsePrototype( com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto.getDefaultInstance(); case 1: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto.getDefaultInstance(); case 2: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.getDefaultInstance(); case 3: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } public static Stub newStub( com.google.protobuf.RpcChannel channel) { return new Stub(channel); } public static final class Stub extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapDaemonProtocol implements Interface { private Stub(com.google.protobuf.RpcChannel channel) { this.channel = channel; } private final com.google.protobuf.RpcChannel channel; public com.google.protobuf.RpcChannel getChannel() { return channel; } public void submitWork( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto> done) { channel.callMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto.getDefaultInstance(), com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto.getDefaultInstance())); } public void sourceStateUpdated( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto> done) { channel.callMethod( getDescriptor().getMethods().get(1), controller, request, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto.getDefaultInstance(), com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto.getDefaultInstance())); } public void queryComplete( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto> done) { channel.callMethod( getDescriptor().getMethods().get(2), controller, request, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.getDefaultInstance(), com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.getDefaultInstance())); } public void terminateFragment( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto> done) { channel.callMethod( getDescriptor().getMethods().get(3), controller, request, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.getDefaultInstance(), com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.getDefaultInstance())); } } public static BlockingInterface newBlockingStub( com.google.protobuf.BlockingRpcChannel channel) { return new BlockingStub(channel); } public interface BlockingInterface { public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto submitWork( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto request) throws com.google.protobuf.ServiceException; public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto sourceStateUpdated( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto request) throws com.google.protobuf.ServiceException; public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto queryComplete( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto request) throws com.google.protobuf.ServiceException; public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto terminateFragment( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto request) throws com.google.protobuf.ServiceException; } private static final class BlockingStub implements BlockingInterface { private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) { this.channel = channel; } private final com.google.protobuf.BlockingRpcChannel channel; public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto submitWork( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto request) throws com.google.protobuf.ServiceException { return (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto) channel.callBlockingMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto.getDefaultInstance()); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto sourceStateUpdated( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto request) throws com.google.protobuf.ServiceException { return (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto) channel.callBlockingMethod( getDescriptor().getMethods().get(1), controller, request, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto.getDefaultInstance()); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto queryComplete( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto request) throws com.google.protobuf.ServiceException { return (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto) channel.callBlockingMethod( getDescriptor().getMethods().get(2), controller, request, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.getDefaultInstance()); } public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto terminateFragment( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto request) throws com.google.protobuf.ServiceException { return (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto) channel.callBlockingMethod( getDescriptor().getMethods().get(3), controller, request, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.getDefaultInstance()); } } // @@protoc_insertion_point(class_scope:LlapDaemonProtocol) } /** * Protobuf service {@code LlapManagementProtocol} */ public static abstract class LlapManagementProtocol implements com.google.protobuf.Service { protected LlapManagementProtocol() {} public interface Interface { /** * <code>rpc getDelegationToken(.GetTokenRequestProto) returns (.GetTokenResponseProto);</code> */ public abstract void getDelegationToken( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto> done); } public static com.google.protobuf.Service newReflectiveService( final Interface impl) { return new LlapManagementProtocol() { @java.lang.Override public void getDelegationToken( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto> done) { impl.getDelegationToken(controller, request, done); } }; } public static com.google.protobuf.BlockingService newReflectiveBlockingService(final BlockingInterface impl) { return new com.google.protobuf.BlockingService() { public final com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final com.google.protobuf.Message callBlockingMethod( com.google.protobuf.Descriptors.MethodDescriptor method, com.google.protobuf.RpcController controller, com.google.protobuf.Message request) throws com.google.protobuf.ServiceException { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callBlockingMethod() given method descriptor for " + "wrong service type."); } switch(method.getIndex()) { case 0: return impl.getDelegationToken(controller, (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto)request); default: throw new java.lang.AssertionError("Can't get here."); } } public final com.google.protobuf.Message getRequestPrototype( com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } public final com.google.protobuf.Message getResponsePrototype( com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } }; } /** * <code>rpc getDelegationToken(.GetTokenRequestProto) returns (.GetTokenResponseProto);</code> */ public abstract void getDelegationToken( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto> done); public static final com.google.protobuf.Descriptors.ServiceDescriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.getDescriptor().getServices().get(1); } public final com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final void callMethod( com.google.protobuf.Descriptors.MethodDescriptor method, com.google.protobuf.RpcController controller, com.google.protobuf.Message request, com.google.protobuf.RpcCallback< com.google.protobuf.Message> done) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callMethod() given method descriptor for wrong " + "service type."); } switch(method.getIndex()) { case 0: this.getDelegationToken(controller, (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto)request, com.google.protobuf.RpcUtil.<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto>specializeCallback( done)); return; default: throw new java.lang.AssertionError("Can't get here."); } } public final com.google.protobuf.Message getRequestPrototype( com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } public final com.google.protobuf.Message getResponsePrototype( com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } public static Stub newStub( com.google.protobuf.RpcChannel channel) { return new Stub(channel); } public static final class Stub extends org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapManagementProtocol implements Interface { private Stub(com.google.protobuf.RpcChannel channel) { this.channel = channel; } private final com.google.protobuf.RpcChannel channel; public com.google.protobuf.RpcChannel getChannel() { return channel; } public void getDelegationToken( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto request, com.google.protobuf.RpcCallback<org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto> done) { channel.callMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto.getDefaultInstance(), com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto.getDefaultInstance())); } } public static BlockingInterface newBlockingStub( com.google.protobuf.BlockingRpcChannel channel) { return new BlockingStub(channel); } public interface BlockingInterface { public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto getDelegationToken( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto request) throws com.google.protobuf.ServiceException; } private static final class BlockingStub implements BlockingInterface { private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) { this.channel = channel; } private final com.google.protobuf.BlockingRpcChannel channel; public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto getDelegationToken( com.google.protobuf.RpcController controller, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto request) throws com.google.protobuf.ServiceException { return (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto) channel.callBlockingMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto.getDefaultInstance()); } } // @@protoc_insertion_point(class_scope:LlapManagementProtocol) } private static com.google.protobuf.Descriptors.Descriptor internal_static_UserPayloadProto_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_UserPayloadProto_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_EntityDescriptorProto_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_EntityDescriptorProto_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_IOSpecProto_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_IOSpecProto_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_GroupInputSpecProto_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_GroupInputSpecProto_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_SignableVertexSpec_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_SignableVertexSpec_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_VertexOrBinary_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_VertexOrBinary_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_FragmentRuntimeInfo_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_FragmentRuntimeInfo_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_QueryIdentifierProto_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_QueryIdentifierProto_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_NotTezEvent_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_NotTezEvent_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_SubmitWorkRequestProto_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_SubmitWorkRequestProto_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_SubmitWorkResponseProto_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_SubmitWorkResponseProto_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_SourceStateUpdatedRequestProto_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_SourceStateUpdatedRequestProto_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_SourceStateUpdatedResponseProto_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_SourceStateUpdatedResponseProto_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_QueryCompleteRequestProto_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_QueryCompleteRequestProto_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_QueryCompleteResponseProto_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_QueryCompleteResponseProto_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_TerminateFragmentRequestProto_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_TerminateFragmentRequestProto_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_TerminateFragmentResponseProto_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_TerminateFragmentResponseProto_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_GetTokenRequestProto_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_GetTokenRequestProto_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_GetTokenResponseProto_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_GetTokenResponseProto_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_LlapOutputSocketInitMessage_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_LlapOutputSocketInitMessage_fieldAccessorTable; public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } private static com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { "\n1llap-common/src/protobuf/LlapDaemonPro" + "tocol.proto\"9\n\020UserPayloadProto\022\024\n\014user_" + "payload\030\001 \001(\014\022\017\n\007version\030\002 \001(\005\"j\n\025Entity" + "DescriptorProto\022\022\n\nclass_name\030\001 \001(\t\022\'\n\014u" + "ser_payload\030\002 \001(\0132\021.UserPayloadProto\022\024\n\014" + "history_text\030\003 \001(\014\"x\n\013IOSpecProto\022\035\n\025con" + "nected_vertex_name\030\001 \001(\t\022-\n\rio_descripto" + "r\030\002 \001(\0132\026.EntityDescriptorProto\022\033\n\023physi" + "cal_edge_count\030\003 \001(\005\"z\n\023GroupInputSpecPr" + "oto\022\022\n\ngroup_name\030\001 \001(\t\022\026\n\016group_vertice", "s\030\002 \003(\t\0227\n\027merged_input_descriptor\030\003 \001(\013" + "2\026.EntityDescriptorProto\"\314\003\n\022SignableVer" + "texSpec\022\014\n\004user\030\001 \001(\t\022\026\n\016signatureKeyId\030" + "\002 \001(\003\022/\n\020query_identifier\030\003 \001(\0132\025.QueryI" + "dentifierProto\022\025\n\rhive_query_id\030\004 \001(\t\022\020\n" + "\010dag_name\030\005 \001(\t\022\023\n\013vertex_name\030\006 \001(\t\022\024\n\014" + "vertex_index\030\007 \001(\005\022\030\n\020token_identifier\030\010" + " \001(\t\0224\n\024processor_descriptor\030\t \001(\0132\026.Ent" + "ityDescriptorProto\022!\n\013input_specs\030\n \003(\0132" + "\014.IOSpecProto\022\"\n\014output_specs\030\013 \003(\0132\014.IO", "SpecProto\0221\n\023grouped_input_specs\030\014 \003(\0132\024" + ".GroupInputSpecProto\022\032\n\022vertex_paralleli" + "sm\030\r \001(\005\022%\n\026is_external_submission\030\016 \001(\010" + ":\005false\"K\n\016VertexOrBinary\022#\n\006vertex\030\001 \001(" + "\0132\023.SignableVertexSpec\022\024\n\014vertexBinary\030\002" + " \001(\014\"\344\001\n\023FragmentRuntimeInfo\022#\n\033num_self" + "_and_upstream_tasks\030\001 \001(\005\022-\n%num_self_an" + "d_upstream_completed_tasks\030\002 \001(\005\022\033\n\023with" + "in_dag_priority\030\003 \001(\005\022\026\n\016dag_start_time\030" + "\004 \001(\003\022 \n\030first_attempt_start_time\030\005 \001(\003\022", "\"\n\032current_attempt_start_time\030\006 \001(\003\"d\n\024Q" + "ueryIdentifierProto\022\035\n\025application_id_st" + "ring\030\001 \001(\t\022\021\n\tdag_index\030\002 \001(\005\022\032\n\022app_att" + "empt_number\030\003 \001(\005\"l\n\013NotTezEvent\022\037\n\027inpu" + "t_event_proto_bytes\030\001 \002(\014\022\023\n\013vertex_name" + "\030\002 \002(\t\022\027\n\017dest_input_name\030\003 \002(\t\022\016\n\006key_i" + "d\030\004 \001(\005\"\330\002\n\026SubmitWorkRequestProto\022\"\n\two" + "rk_spec\030\001 \001(\0132\017.VertexOrBinary\022\033\n\023work_s" + "pec_signature\030\002 \001(\014\022\027\n\017fragment_number\030\003" + " \001(\005\022\026\n\016attempt_number\030\004 \001(\005\022\033\n\023containe", "r_id_string\030\005 \001(\t\022\017\n\007am_host\030\006 \001(\t\022\017\n\007am" + "_port\030\007 \001(\005\022\032\n\022credentials_binary\030\010 \001(\014\022" + "3\n\025fragment_runtime_info\030\t \001(\0132\024.Fragmen" + "tRuntimeInfo\022\033\n\023initial_event_bytes\030\n \001(" + "\014\022\037\n\027initial_event_signature\030\013 \001(\014\"b\n\027Su" + "bmitWorkResponseProto\022/\n\020submission_stat" + "e\030\001 \001(\0162\025.SubmissionStateProto\022\026\n\016unique" + "_node_id\030\002 \001(\t\"\205\001\n\036SourceStateUpdatedReq" + "uestProto\022/\n\020query_identifier\030\001 \001(\0132\025.Qu" + "eryIdentifierProto\022\020\n\010src_name\030\002 \001(\t\022 \n\005", "state\030\003 \001(\0162\021.SourceStateProto\"!\n\037Source" + "StateUpdatedResponseProto\"e\n\031QueryComple" + "teRequestProto\022/\n\020query_identifier\030\001 \001(\013" + "2\025.QueryIdentifierProto\022\027\n\014delete_delay\030" + "\002 \001(\003:\0010\"\034\n\032QueryCompleteResponseProto\"t" + "\n\035TerminateFragmentRequestProto\022/\n\020query" + "_identifier\030\001 \001(\0132\025.QueryIdentifierProto" + "\022\"\n\032fragment_identifier_string\030\002 \001(\t\" \n\036" + "TerminateFragmentResponseProto\"&\n\024GetTok" + "enRequestProto\022\016\n\006app_id\030\001 \001(\t\"&\n\025GetTok", "enResponseProto\022\r\n\005token\030\001 \001(\014\"A\n\033LlapOu" + "tputSocketInitMessage\022\023\n\013fragment_id\030\001 \002" + "(\t\022\r\n\005token\030\002 \001(\014*2\n\020SourceStateProto\022\017\n" + "\013S_SUCCEEDED\020\001\022\r\n\tS_RUNNING\020\002*E\n\024Submiss" + "ionStateProto\022\014\n\010ACCEPTED\020\001\022\014\n\010REJECTED\020" + "\002\022\021\n\rEVICTED_OTHER\020\0032\316\002\n\022LlapDaemonProto" + "col\022?\n\nsubmitWork\022\027.SubmitWorkRequestPro" + "to\032\030.SubmitWorkResponseProto\022W\n\022sourceSt" + "ateUpdated\022\037.SourceStateUpdatedRequestPr" + "oto\032 .SourceStateUpdatedResponseProto\022H\n", "\rqueryComplete\022\032.QueryCompleteRequestPro" + "to\032\033.QueryCompleteResponseProto\022T\n\021termi" + "nateFragment\022\036.TerminateFragmentRequestP" + "roto\032\037.TerminateFragmentResponseProto2]\n" + "\026LlapManagementProtocol\022C\n\022getDelegation" + "Token\022\025.GetTokenRequestProto\032\026.GetTokenR" + "esponseProtoBH\n&org.apache.hadoop.hive.l" + "lap.daemon.rpcB\030LlapDaemonProtocolProtos" + "\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { public com.google.protobuf.ExtensionRegistry assignDescriptors( com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_UserPayloadProto_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_UserPayloadProto_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_UserPayloadProto_descriptor, new java.lang.String[] { "UserPayload", "Version", }); internal_static_EntityDescriptorProto_descriptor = getDescriptor().getMessageTypes().get(1); internal_static_EntityDescriptorProto_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_EntityDescriptorProto_descriptor, new java.lang.String[] { "ClassName", "UserPayload", "HistoryText", }); internal_static_IOSpecProto_descriptor = getDescriptor().getMessageTypes().get(2); internal_static_IOSpecProto_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_IOSpecProto_descriptor, new java.lang.String[] { "ConnectedVertexName", "IoDescriptor", "PhysicalEdgeCount", }); internal_static_GroupInputSpecProto_descriptor = getDescriptor().getMessageTypes().get(3); internal_static_GroupInputSpecProto_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_GroupInputSpecProto_descriptor, new java.lang.String[] { "GroupName", "GroupVertices", "MergedInputDescriptor", }); internal_static_SignableVertexSpec_descriptor = getDescriptor().getMessageTypes().get(4); internal_static_SignableVertexSpec_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_SignableVertexSpec_descriptor, new java.lang.String[] { "User", "SignatureKeyId", "QueryIdentifier", "HiveQueryId", "DagName", "VertexName", "VertexIndex", "TokenIdentifier", "ProcessorDescriptor", "InputSpecs", "OutputSpecs", "GroupedInputSpecs", "VertexParallelism", "IsExternalSubmission", }); internal_static_VertexOrBinary_descriptor = getDescriptor().getMessageTypes().get(5); internal_static_VertexOrBinary_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_VertexOrBinary_descriptor, new java.lang.String[] { "Vertex", "VertexBinary", }); internal_static_FragmentRuntimeInfo_descriptor = getDescriptor().getMessageTypes().get(6); internal_static_FragmentRuntimeInfo_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_FragmentRuntimeInfo_descriptor, new java.lang.String[] { "NumSelfAndUpstreamTasks", "NumSelfAndUpstreamCompletedTasks", "WithinDagPriority", "DagStartTime", "FirstAttemptStartTime", "CurrentAttemptStartTime", }); internal_static_QueryIdentifierProto_descriptor = getDescriptor().getMessageTypes().get(7); internal_static_QueryIdentifierProto_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_QueryIdentifierProto_descriptor, new java.lang.String[] { "ApplicationIdString", "DagIndex", "AppAttemptNumber", }); internal_static_NotTezEvent_descriptor = getDescriptor().getMessageTypes().get(8); internal_static_NotTezEvent_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_NotTezEvent_descriptor, new java.lang.String[] { "InputEventProtoBytes", "VertexName", "DestInputName", "KeyId", }); internal_static_SubmitWorkRequestProto_descriptor = getDescriptor().getMessageTypes().get(9); internal_static_SubmitWorkRequestProto_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_SubmitWorkRequestProto_descriptor, new java.lang.String[] { "WorkSpec", "WorkSpecSignature", "FragmentNumber", "AttemptNumber", "ContainerIdString", "AmHost", "AmPort", "CredentialsBinary", "FragmentRuntimeInfo", "InitialEventBytes", "InitialEventSignature", }); internal_static_SubmitWorkResponseProto_descriptor = getDescriptor().getMessageTypes().get(10); internal_static_SubmitWorkResponseProto_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_SubmitWorkResponseProto_descriptor, new java.lang.String[] { "SubmissionState", "UniqueNodeId", }); internal_static_SourceStateUpdatedRequestProto_descriptor = getDescriptor().getMessageTypes().get(11); internal_static_SourceStateUpdatedRequestProto_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_SourceStateUpdatedRequestProto_descriptor, new java.lang.String[] { "QueryIdentifier", "SrcName", "State", }); internal_static_SourceStateUpdatedResponseProto_descriptor = getDescriptor().getMessageTypes().get(12); internal_static_SourceStateUpdatedResponseProto_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_SourceStateUpdatedResponseProto_descriptor, new java.lang.String[] { }); internal_static_QueryCompleteRequestProto_descriptor = getDescriptor().getMessageTypes().get(13); internal_static_QueryCompleteRequestProto_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_QueryCompleteRequestProto_descriptor, new java.lang.String[] { "QueryIdentifier", "DeleteDelay", }); internal_static_QueryCompleteResponseProto_descriptor = getDescriptor().getMessageTypes().get(14); internal_static_QueryCompleteResponseProto_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_QueryCompleteResponseProto_descriptor, new java.lang.String[] { }); internal_static_TerminateFragmentRequestProto_descriptor = getDescriptor().getMessageTypes().get(15); internal_static_TerminateFragmentRequestProto_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_TerminateFragmentRequestProto_descriptor, new java.lang.String[] { "QueryIdentifier", "FragmentIdentifierString", }); internal_static_TerminateFragmentResponseProto_descriptor = getDescriptor().getMessageTypes().get(16); internal_static_TerminateFragmentResponseProto_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_TerminateFragmentResponseProto_descriptor, new java.lang.String[] { }); internal_static_GetTokenRequestProto_descriptor = getDescriptor().getMessageTypes().get(17); internal_static_GetTokenRequestProto_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_GetTokenRequestProto_descriptor, new java.lang.String[] { "AppId", }); internal_static_GetTokenResponseProto_descriptor = getDescriptor().getMessageTypes().get(18); internal_static_GetTokenResponseProto_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_GetTokenResponseProto_descriptor, new java.lang.String[] { "Token", }); internal_static_LlapOutputSocketInitMessage_descriptor = getDescriptor().getMessageTypes().get(19); internal_static_LlapOutputSocketInitMessage_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_LlapOutputSocketInitMessage_descriptor, new java.lang.String[] { "FragmentId", "Token", }); return null; } }; com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } // @@protoc_insertion_point(outer_class_scope) }