http://git-wip-us.apache.org/repos/asf/hive/blob/9886414b/llap-server/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java ---------------------------------------------------------------------- diff --git a/llap-server/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java b/llap-server/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java deleted file mode 100644 index 4ab7b32..0000000 --- a/llap-server/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java +++ /dev/null @@ -1,14553 +0,0 @@ -// Generated by the protocol buffer compiler. DO NOT EDIT! -// source: LlapDaemonProtocol.proto - -package org.apache.hadoop.hive.llap.daemon.rpc; - -public final class LlapDaemonProtocolProtos { - private LlapDaemonProtocolProtos() {} - public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { - } - /** - * Protobuf enum {@code SourceStateProto} - */ - public enum SourceStateProto - implements com.google.protobuf.ProtocolMessageEnum { - /** - * <code>S_SUCCEEDED = 1;</code> - */ - S_SUCCEEDED(0, 1), - /** - * <code>S_RUNNING = 2;</code> - */ - S_RUNNING(1, 2), - ; - - /** - * <code>S_SUCCEEDED = 1;</code> - */ - public static final int S_SUCCEEDED_VALUE = 1; - /** - * <code>S_RUNNING = 2;</code> - */ - public static final int S_RUNNING_VALUE = 2; - - - public final int getNumber() { return value; } - - public static SourceStateProto valueOf(int value) { - switch (value) { - case 1: return S_SUCCEEDED; - case 2: return S_RUNNING; - default: return null; - } - } - - public static com.google.protobuf.Internal.EnumLiteMap<SourceStateProto> - internalGetValueMap() { - return internalValueMap; - } - private static com.google.protobuf.Internal.EnumLiteMap<SourceStateProto> - internalValueMap = - new com.google.protobuf.Internal.EnumLiteMap<SourceStateProto>() { - public SourceStateProto findValueByNumber(int number) { - return SourceStateProto.valueOf(number); - } - }; - - public final com.google.protobuf.Descriptors.EnumValueDescriptor - getValueDescriptor() { - return getDescriptor().getValues().get(index); - } - public final com.google.protobuf.Descriptors.EnumDescriptor - getDescriptorForType() { - return getDescriptor(); - } - public static final com.google.protobuf.Descriptors.EnumDescriptor - getDescriptor() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.getDescriptor().getEnumTypes().get(0); - } - - private static final SourceStateProto[] VALUES = values(); - - public static SourceStateProto valueOf( - com.google.protobuf.Descriptors.EnumValueDescriptor desc) { - if (desc.getType() != getDescriptor()) { - throw new java.lang.IllegalArgumentException( - "EnumValueDescriptor is not for this type."); - } - return VALUES[desc.getIndex()]; - } - - private final int index; - private final int value; - - private SourceStateProto(int index, int value) { - this.index = index; - this.value = value; - } - - // @@protoc_insertion_point(enum_scope:SourceStateProto) - } - - /** - * Protobuf enum {@code SubmissionStateProto} - */ - public enum SubmissionStateProto - implements com.google.protobuf.ProtocolMessageEnum { - /** - * <code>ACCEPTED = 1;</code> - */ - ACCEPTED(0, 1), - /** - * <code>REJECTED = 2;</code> - */ - REJECTED(1, 2), - /** - * <code>EVICTED_OTHER = 3;</code> - */ - EVICTED_OTHER(2, 3), - ; - - /** - * <code>ACCEPTED = 1;</code> - */ - public static final int ACCEPTED_VALUE = 1; - /** - * <code>REJECTED = 2;</code> - */ - public static final int REJECTED_VALUE = 2; - /** - * <code>EVICTED_OTHER = 3;</code> - */ - public static final int EVICTED_OTHER_VALUE = 3; - - - public final int getNumber() { return value; } - - public static SubmissionStateProto valueOf(int value) { - switch (value) { - case 1: return ACCEPTED; - case 2: return REJECTED; - case 3: return EVICTED_OTHER; - default: return null; - } - } - - public static com.google.protobuf.Internal.EnumLiteMap<SubmissionStateProto> - internalGetValueMap() { - return internalValueMap; - } - private static com.google.protobuf.Internal.EnumLiteMap<SubmissionStateProto> - internalValueMap = - new com.google.protobuf.Internal.EnumLiteMap<SubmissionStateProto>() { - public SubmissionStateProto findValueByNumber(int number) { - return SubmissionStateProto.valueOf(number); - } - }; - - public final com.google.protobuf.Descriptors.EnumValueDescriptor - getValueDescriptor() { - return getDescriptor().getValues().get(index); - } - public final com.google.protobuf.Descriptors.EnumDescriptor - getDescriptorForType() { - return getDescriptor(); - } - public static final com.google.protobuf.Descriptors.EnumDescriptor - getDescriptor() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.getDescriptor().getEnumTypes().get(1); - } - - private static final SubmissionStateProto[] VALUES = values(); - - public static SubmissionStateProto valueOf( - com.google.protobuf.Descriptors.EnumValueDescriptor desc) { - if (desc.getType() != getDescriptor()) { - throw new java.lang.IllegalArgumentException( - "EnumValueDescriptor is not for this type."); - } - return VALUES[desc.getIndex()]; - } - - private final int index; - private final int value; - - private SubmissionStateProto(int index, int value) { - this.index = index; - this.value = value; - } - - // @@protoc_insertion_point(enum_scope:SubmissionStateProto) - } - - public interface UserPayloadProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { - - // optional bytes user_payload = 1; - /** - * <code>optional bytes user_payload = 1;</code> - */ - boolean hasUserPayload(); - /** - * <code>optional bytes user_payload = 1;</code> - */ - com.google.protobuf.ByteString getUserPayload(); - - // optional int32 version = 2; - /** - * <code>optional int32 version = 2;</code> - */ - boolean hasVersion(); - /** - * <code>optional int32 version = 2;</code> - */ - int getVersion(); - } - /** - * Protobuf type {@code UserPayloadProto} - */ - public static final class UserPayloadProto extends - com.google.protobuf.GeneratedMessage - implements UserPayloadProtoOrBuilder { - // Use UserPayloadProto.newBuilder() to construct. - private UserPayloadProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { - super(builder); - this.unknownFields = builder.getUnknownFields(); - } - private UserPayloadProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final UserPayloadProto defaultInstance; - public static UserPayloadProto getDefaultInstance() { - return defaultInstance; - } - - public UserPayloadProto getDefaultInstanceForType() { - return defaultInstance; - } - - private final com.google.protobuf.UnknownFieldSet unknownFields; - @java.lang.Override - public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { - return this.unknownFields; - } - private UserPayloadProto( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); - int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); - try { - boolean done = false; - while (!done) { - int tag = input.readTag(); - switch (tag) { - case 0: - done = true; - break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } - case 10: { - bitField0_ |= 0x00000001; - userPayload_ = input.readBytes(); - break; - } - case 16: { - bitField0_ |= 0x00000002; - version_ = input.readInt32(); - break; - } - } - } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { - throw e.setUnfinishedMessage(this); - } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); - } finally { - this.unknownFields = unknownFields.build(); - makeExtensionsImmutable(); - } - } - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder.class); - } - - public static com.google.protobuf.Parser<UserPayloadProto> PARSER = - new com.google.protobuf.AbstractParser<UserPayloadProto>() { - public UserPayloadProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new UserPayloadProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser<UserPayloadProto> getParserForType() { - return PARSER; - } - - private int bitField0_; - // optional bytes user_payload = 1; - public static final int USER_PAYLOAD_FIELD_NUMBER = 1; - private com.google.protobuf.ByteString userPayload_; - /** - * <code>optional bytes user_payload = 1;</code> - */ - public boolean hasUserPayload() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * <code>optional bytes user_payload = 1;</code> - */ - public com.google.protobuf.ByteString getUserPayload() { - return userPayload_; - } - - // optional int32 version = 2; - public static final int VERSION_FIELD_NUMBER = 2; - private int version_; - /** - * <code>optional int32 version = 2;</code> - */ - public boolean hasVersion() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * <code>optional int32 version = 2;</code> - */ - public int getVersion() { - return version_; - } - - private void initFields() { - userPayload_ = com.google.protobuf.ByteString.EMPTY; - version_ = 0; - } - private byte memoizedIsInitialized = -1; - public final boolean isInitialized() { - byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; - - memoizedIsInitialized = 1; - return true; - } - - public void writeTo(com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, userPayload_); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeInt32(2, version_); - } - getUnknownFields().writeTo(output); - } - - private int memoizedSerializedSize = -1; - public int getSerializedSize() { - int size = memoizedSerializedSize; - if (size != -1) return size; - - size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, userPayload_); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeInt32Size(2, version_); - } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; - return size; - } - - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override - public boolean equals(final java.lang.Object obj) { - if (obj == this) { - return true; - } - if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto)) { - return super.equals(obj); - } - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto) obj; - - boolean result = true; - result = result && (hasUserPayload() == other.hasUserPayload()); - if (hasUserPayload()) { - result = result && getUserPayload() - .equals(other.getUserPayload()); - } - result = result && (hasVersion() == other.hasVersion()); - if (hasVersion()) { - result = result && (getVersion() - == other.getVersion()); - } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; - } - - private int memoizedHashCode = 0; - @java.lang.Override - public int hashCode() { - if (memoizedHashCode != 0) { - return memoizedHashCode; - } - int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasUserPayload()) { - hash = (37 * hash) + USER_PAYLOAD_FIELD_NUMBER; - hash = (53 * hash) + getUserPayload().hashCode(); - } - if (hasVersion()) { - hash = (37 * hash) + VERSION_FIELD_NUMBER; - hash = (53 * hash) + getVersion(); - } - hash = (29 * hash) + getUnknownFields().hashCode(); - memoizedHashCode = hash; - return hash; - } - - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom( - byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(java.io.InputStream input) - throws java.io.IOException { - return PARSER.parseFrom(input); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseDelimitedFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom( - com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return PARSER.parseFrom(input); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); - } - - public static Builder newBuilder() { return Builder.create(); } - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto prototype) { - return newBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { return newBuilder(this); } - - @java.lang.Override - protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - Builder builder = new Builder(parent); - return builder; - } - /** - * Protobuf type {@code UserPayloadProto} - */ - public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder<Builder> - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder.class); - } - - // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.newBuilder() - private Builder() { - maybeForceBuilderInitialization(); - } - - private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - super(parent); - maybeForceBuilderInitialization(); - } - private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { - } - } - private static Builder create() { - return new Builder(); - } - - public Builder clear() { - super.clear(); - userPayload_ = com.google.protobuf.ByteString.EMPTY; - bitField0_ = (bitField0_ & ~0x00000001); - version_ = 0; - bitField0_ = (bitField0_ & ~0x00000002); - return this; - } - - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - - public com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_descriptor; - } - - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getDefaultInstanceForType() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance(); - } - - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto build() { - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException(result); - } - return result; - } - - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto buildPartial() { - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto(this); - int from_bitField0_ = bitField0_; - int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { - to_bitField0_ |= 0x00000001; - } - result.userPayload_ = userPayload_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { - to_bitField0_ |= 0x00000002; - } - result.version_ = version_; - result.bitField0_ = to_bitField0_; - onBuilt(); - return result; - } - - public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto) { - return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto other) { - if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance()) return this; - if (other.hasUserPayload()) { - setUserPayload(other.getUserPayload()); - } - if (other.hasVersion()) { - setVersion(other.getVersion()); - } - this.mergeUnknownFields(other.getUnknownFields()); - return this; - } - - public final boolean isInitialized() { - return true; - } - - public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parsedMessage = null; - try { - parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto) e.getUnfinishedMessage(); - throw e; - } finally { - if (parsedMessage != null) { - mergeFrom(parsedMessage); - } - } - return this; - } - private int bitField0_; - - // optional bytes user_payload = 1; - private com.google.protobuf.ByteString userPayload_ = com.google.protobuf.ByteString.EMPTY; - /** - * <code>optional bytes user_payload = 1;</code> - */ - public boolean hasUserPayload() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * <code>optional bytes user_payload = 1;</code> - */ - public com.google.protobuf.ByteString getUserPayload() { - return userPayload_; - } - /** - * <code>optional bytes user_payload = 1;</code> - */ - public Builder setUserPayload(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000001; - userPayload_ = value; - onChanged(); - return this; - } - /** - * <code>optional bytes user_payload = 1;</code> - */ - public Builder clearUserPayload() { - bitField0_ = (bitField0_ & ~0x00000001); - userPayload_ = getDefaultInstance().getUserPayload(); - onChanged(); - return this; - } - - // optional int32 version = 2; - private int version_ ; - /** - * <code>optional int32 version = 2;</code> - */ - public boolean hasVersion() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * <code>optional int32 version = 2;</code> - */ - public int getVersion() { - return version_; - } - /** - * <code>optional int32 version = 2;</code> - */ - public Builder setVersion(int value) { - bitField0_ |= 0x00000002; - version_ = value; - onChanged(); - return this; - } - /** - * <code>optional int32 version = 2;</code> - */ - public Builder clearVersion() { - bitField0_ = (bitField0_ & ~0x00000002); - version_ = 0; - onChanged(); - return this; - } - - // @@protoc_insertion_point(builder_scope:UserPayloadProto) - } - - static { - defaultInstance = new UserPayloadProto(true); - defaultInstance.initFields(); - } - - // @@protoc_insertion_point(class_scope:UserPayloadProto) - } - - public interface EntityDescriptorProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { - - // optional string class_name = 1; - /** - * <code>optional string class_name = 1;</code> - */ - boolean hasClassName(); - /** - * <code>optional string class_name = 1;</code> - */ - java.lang.String getClassName(); - /** - * <code>optional string class_name = 1;</code> - */ - com.google.protobuf.ByteString - getClassNameBytes(); - - // optional .UserPayloadProto user_payload = 2; - /** - * <code>optional .UserPayloadProto user_payload = 2;</code> - */ - boolean hasUserPayload(); - /** - * <code>optional .UserPayloadProto user_payload = 2;</code> - */ - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getUserPayload(); - /** - * <code>optional .UserPayloadProto user_payload = 2;</code> - */ - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder getUserPayloadOrBuilder(); - - // optional bytes history_text = 3; - /** - * <code>optional bytes history_text = 3;</code> - */ - boolean hasHistoryText(); - /** - * <code>optional bytes history_text = 3;</code> - */ - com.google.protobuf.ByteString getHistoryText(); - } - /** - * Protobuf type {@code EntityDescriptorProto} - */ - public static final class EntityDescriptorProto extends - com.google.protobuf.GeneratedMessage - implements EntityDescriptorProtoOrBuilder { - // Use EntityDescriptorProto.newBuilder() to construct. - private EntityDescriptorProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { - super(builder); - this.unknownFields = builder.getUnknownFields(); - } - private EntityDescriptorProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final EntityDescriptorProto defaultInstance; - public static EntityDescriptorProto getDefaultInstance() { - return defaultInstance; - } - - public EntityDescriptorProto getDefaultInstanceForType() { - return defaultInstance; - } - - private final com.google.protobuf.UnknownFieldSet unknownFields; - @java.lang.Override - public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { - return this.unknownFields; - } - private EntityDescriptorProto( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); - int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); - try { - boolean done = false; - while (!done) { - int tag = input.readTag(); - switch (tag) { - case 0: - done = true; - break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } - case 10: { - bitField0_ |= 0x00000001; - className_ = input.readBytes(); - break; - } - case 18: { - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder subBuilder = null; - if (((bitField0_ & 0x00000002) == 0x00000002)) { - subBuilder = userPayload_.toBuilder(); - } - userPayload_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.PARSER, extensionRegistry); - if (subBuilder != null) { - subBuilder.mergeFrom(userPayload_); - userPayload_ = subBuilder.buildPartial(); - } - bitField0_ |= 0x00000002; - break; - } - case 26: { - bitField0_ |= 0x00000004; - historyText_ = input.readBytes(); - break; - } - } - } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { - throw e.setUnfinishedMessage(this); - } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); - } finally { - this.unknownFields = unknownFields.build(); - makeExtensionsImmutable(); - } - } - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder.class); - } - - public static com.google.protobuf.Parser<EntityDescriptorProto> PARSER = - new com.google.protobuf.AbstractParser<EntityDescriptorProto>() { - public EntityDescriptorProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new EntityDescriptorProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser<EntityDescriptorProto> getParserForType() { - return PARSER; - } - - private int bitField0_; - // optional string class_name = 1; - public static final int CLASS_NAME_FIELD_NUMBER = 1; - private java.lang.Object className_; - /** - * <code>optional string class_name = 1;</code> - */ - public boolean hasClassName() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * <code>optional string class_name = 1;</code> - */ - public java.lang.String getClassName() { - java.lang.Object ref = className_; - if (ref instanceof java.lang.String) { - return (java.lang.String) ref; - } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; - java.lang.String s = bs.toStringUtf8(); - if (bs.isValidUtf8()) { - className_ = s; - } - return s; - } - } - /** - * <code>optional string class_name = 1;</code> - */ - public com.google.protobuf.ByteString - getClassNameBytes() { - java.lang.Object ref = className_; - if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - className_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - - // optional .UserPayloadProto user_payload = 2; - public static final int USER_PAYLOAD_FIELD_NUMBER = 2; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto userPayload_; - /** - * <code>optional .UserPayloadProto user_payload = 2;</code> - */ - public boolean hasUserPayload() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * <code>optional .UserPayloadProto user_payload = 2;</code> - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getUserPayload() { - return userPayload_; - } - /** - * <code>optional .UserPayloadProto user_payload = 2;</code> - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder getUserPayloadOrBuilder() { - return userPayload_; - } - - // optional bytes history_text = 3; - public static final int HISTORY_TEXT_FIELD_NUMBER = 3; - private com.google.protobuf.ByteString historyText_; - /** - * <code>optional bytes history_text = 3;</code> - */ - public boolean hasHistoryText() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - /** - * <code>optional bytes history_text = 3;</code> - */ - public com.google.protobuf.ByteString getHistoryText() { - return historyText_; - } - - private void initFields() { - className_ = ""; - userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance(); - historyText_ = com.google.protobuf.ByteString.EMPTY; - } - private byte memoizedIsInitialized = -1; - public final boolean isInitialized() { - byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; - - memoizedIsInitialized = 1; - return true; - } - - public void writeTo(com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getClassNameBytes()); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeMessage(2, userPayload_); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeBytes(3, historyText_); - } - getUnknownFields().writeTo(output); - } - - private int memoizedSerializedSize = -1; - public int getSerializedSize() { - int size = memoizedSerializedSize; - if (size != -1) return size; - - size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getClassNameBytes()); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(2, userPayload_); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(3, historyText_); - } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; - return size; - } - - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override - public boolean equals(final java.lang.Object obj) { - if (obj == this) { - return true; - } - if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto)) { - return super.equals(obj); - } - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto) obj; - - boolean result = true; - result = result && (hasClassName() == other.hasClassName()); - if (hasClassName()) { - result = result && getClassName() - .equals(other.getClassName()); - } - result = result && (hasUserPayload() == other.hasUserPayload()); - if (hasUserPayload()) { - result = result && getUserPayload() - .equals(other.getUserPayload()); - } - result = result && (hasHistoryText() == other.hasHistoryText()); - if (hasHistoryText()) { - result = result && getHistoryText() - .equals(other.getHistoryText()); - } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; - } - - private int memoizedHashCode = 0; - @java.lang.Override - public int hashCode() { - if (memoizedHashCode != 0) { - return memoizedHashCode; - } - int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasClassName()) { - hash = (37 * hash) + CLASS_NAME_FIELD_NUMBER; - hash = (53 * hash) + getClassName().hashCode(); - } - if (hasUserPayload()) { - hash = (37 * hash) + USER_PAYLOAD_FIELD_NUMBER; - hash = (53 * hash) + getUserPayload().hashCode(); - } - if (hasHistoryText()) { - hash = (37 * hash) + HISTORY_TEXT_FIELD_NUMBER; - hash = (53 * hash) + getHistoryText().hashCode(); - } - hash = (29 * hash) + getUnknownFields().hashCode(); - memoizedHashCode = hash; - return hash; - } - - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom( - byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(java.io.InputStream input) - throws java.io.IOException { - return PARSER.parseFrom(input); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseDelimitedFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom( - com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return PARSER.parseFrom(input); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); - } - - public static Builder newBuilder() { return Builder.create(); } - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto prototype) { - return newBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { return newBuilder(this); } - - @java.lang.Override - protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - Builder builder = new Builder(parent); - return builder; - } - /** - * Protobuf type {@code EntityDescriptorProto} - */ - public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder<Builder> - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder.class); - } - - // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.newBuilder() - private Builder() { - maybeForceBuilderInitialization(); - } - - private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - super(parent); - maybeForceBuilderInitialization(); - } - private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { - getUserPayloadFieldBuilder(); - } - } - private static Builder create() { - return new Builder(); - } - - public Builder clear() { - super.clear(); - className_ = ""; - bitField0_ = (bitField0_ & ~0x00000001); - if (userPayloadBuilder_ == null) { - userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance(); - } else { - userPayloadBuilder_.clear(); - } - bitField0_ = (bitField0_ & ~0x00000002); - historyText_ = com.google.protobuf.ByteString.EMPTY; - bitField0_ = (bitField0_ & ~0x00000004); - return this; - } - - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - - public com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_descriptor; - } - - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getDefaultInstanceForType() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); - } - - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto build() { - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException(result); - } - return result; - } - - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto buildPartial() { - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto(this); - int from_bitField0_ = bitField0_; - int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { - to_bitField0_ |= 0x00000001; - } - result.className_ = className_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { - to_bitField0_ |= 0x00000002; - } - if (userPayloadBuilder_ == null) { - result.userPayload_ = userPayload_; - } else { - result.userPayload_ = userPayloadBuilder_.build(); - } - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { - to_bitField0_ |= 0x00000004; - } - result.historyText_ = historyText_; - result.bitField0_ = to_bitField0_; - onBuilt(); - return result; - } - - public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto) { - return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto other) { - if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) return this; - if (other.hasClassName()) { - bitField0_ |= 0x00000001; - className_ = other.className_; - onChanged(); - } - if (other.hasUserPayload()) { - mergeUserPayload(other.getUserPayload()); - } - if (other.hasHistoryText()) { - setHistoryText(other.getHistoryText()); - } - this.mergeUnknownFields(other.getUnknownFields()); - return this; - } - - public final boolean isInitialized() { - return true; - } - - public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parsedMessage = null; - try { - parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto) e.getUnfinishedMessage(); - throw e; - } finally { - if (parsedMessage != null) { - mergeFrom(parsedMessage); - } - } - return this; - } - private int bitField0_; - - // optional string class_name = 1; - private java.lang.Object className_ = ""; - /** - * <code>optional string class_name = 1;</code> - */ - public boolean hasClassName() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * <code>optional string class_name = 1;</code> - */ - public java.lang.String getClassName() { - java.lang.Object ref = className_; - if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - className_ = s; - return s; - } else { - return (java.lang.String) ref; - } - } - /** - * <code>optional string class_name = 1;</code> - */ - public com.google.protobuf.ByteString - getClassNameBytes() { - java.lang.Object ref = className_; - if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - className_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - /** - * <code>optional string class_name = 1;</code> - */ - public Builder setClassName( - java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000001; - className_ = value; - onChanged(); - return this; - } - /** - * <code>optional string class_name = 1;</code> - */ - public Builder clearClassName() { - bitField0_ = (bitField0_ & ~0x00000001); - className_ = getDefaultInstance().getClassName(); - onChanged(); - return this; - } - /** - * <code>optional string class_name = 1;</code> - */ - public Builder setClassNameBytes( - com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000001; - className_ = value; - onChanged(); - return this; - } - - // optional .UserPayloadProto user_payload = 2; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder> userPayloadBuilder_; - /** - * <code>optional .UserPayloadProto user_payload = 2;</code> - */ - public boolean hasUserPayload() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * <code>optional .UserPayloadProto user_payload = 2;</code> - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getUserPayload() { - if (userPayloadBuilder_ == null) { - return userPayload_; - } else { - return userPayloadBuilder_.getMessage(); - } - } - /** - * <code>optional .UserPayloadProto user_payload = 2;</code> - */ - public Builder setUserPayload(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto value) { - if (userPayloadBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - userPayload_ = value; - onChanged(); - } else { - userPayloadBuilder_.setMessage(value); - } - bitField0_ |= 0x00000002; - return this; - } - /** - * <code>optional .UserPayloadProto user_payload = 2;</code> - */ - public Builder setUserPayload( - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder builderForValue) { - if (userPayloadBuilder_ == null) { - userPayload_ = builderForValue.build(); - onChanged(); - } else { - userPayloadBuilder_.setMessage(builderForValue.build()); - } - bitField0_ |= 0x00000002; - return this; - } - /** - * <code>optional .UserPayloadProto user_payload = 2;</code> - */ - public Builder mergeUserPayload(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto value) { - if (userPayloadBuilder_ == null) { - if (((bitField0_ & 0x00000002) == 0x00000002) && - userPayload_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance()) { - userPayload_ = - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.newBuilder(userPayload_).mergeFrom(value).buildPartial(); - } else { - userPayload_ = value; - } - onChanged(); - } else { - userPayloadBuilder_.mergeFrom(value); - } - bitField0_ |= 0x00000002; - return this; - } - /** - * <code>optional .UserPayloadProto user_payload = 2;</code> - */ - public Builder clearUserPayload() { - if (userPayloadBuilder_ == null) { - userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance(); - onChanged(); - } else { - userPayloadBuilder_.clear(); - } - bitField0_ = (bitField0_ & ~0x00000002); - return this; - } - /** - * <code>optional .UserPayloadProto user_payload = 2;</code> - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder getUserPayloadBuilder() { - bitField0_ |= 0x00000002; - onChanged(); - return getUserPayloadFieldBuilder().getBuilder(); - } - /** - * <code>optional .UserPayloadProto user_payload = 2;</code> - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder getUserPayloadOrBuilder() { - if (userPayloadBuilder_ != null) { - return userPayloadBuilder_.getMessageOrBuilder(); - } else { - return userPayload_; - } - } - /** - * <code>optional .UserPayloadProto user_payload = 2;</code> - */ - private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder> - getUserPayloadFieldBuilder() { - if (userPayloadBuilder_ == null) { - userPayloadBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder>( - userPayload_, - getParentForChildren(), - isClean()); - userPayload_ = null; - } - return userPayloadBuilder_; - } - - // optional bytes history_text = 3; - private com.google.protobuf.ByteString historyText_ = com.google.protobuf.ByteString.EMPTY; - /** - * <code>optional bytes history_text = 3;</code> - */ - public boolean hasHistoryText() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - /** - * <code>optional bytes history_text = 3;</code> - */ - public com.google.protobuf.ByteString getHistoryText() { - return historyText_; - } - /** - * <code>optional bytes history_text = 3;</code> - */ - public Builder setHistoryText(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000004; - historyText_ = value; - onChanged(); - return this; - } - /** - * <code>optional bytes history_text = 3;</code> - */ - public Builder clearHistoryText() { - bitField0_ = (bitField0_ & ~0x00000004); - historyText_ = getDefaultInstance().getHistoryText(); - onChanged(); - return this; - } - - // @@protoc_insertion_point(builder_scope:EntityDescriptorProto) - } - - static { - defaultInstance = new EntityDescriptorProto(true); - defaultInstance.initFields(); - } - - // @@protoc_insertion_point(class_scope:EntityDescriptorProto) - } - - public interface IOSpecProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { - - // optional string connected_vertex_name = 1; - /** - * <code>optional string connected_vertex_name = 1;</code> - */ - boolean hasConnectedVertexName(); - /** - * <code>optional string connected_vertex_name = 1;</code> - */ - java.lang.String getConnectedVertexName(); - /** - * <code>optional string connected_vertex_name = 1;</code> - */ - com.google.protobuf.ByteString - getConnectedVertexNameBytes(); - - // optional .EntityDescriptorProto io_descriptor = 2; - /** - * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> - */ - boolean hasIoDescriptor(); - /** - * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> - */ - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getIoDescriptor(); - /** - * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> - */ - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getIoDescriptorOrBuilder(); - - // optional int32 physical_edge_count = 3; - /** - * <code>optional int32 physical_edge_count = 3;</code> - */ - boolean hasPhysicalEdgeCount(); - /** - * <code>optional int32 physical_edge_count = 3;</code> - */ - int getPhysicalEdgeCount(); - } - /** - * Protobuf type {@code IOSpecProto} - */ - public static final class IOSpecProto extends - com.google.protobuf.GeneratedMessage - implements IOSpecProtoOrBuilder { - // Use IOSpecProto.newBuilder() to construct. - private IOSpecProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) { - super(builder); - this.unknownFields = builder.getUnknownFields(); - } - private IOSpecProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final IOSpecProto defaultInstance; - public static IOSpecProto getDefaultInstance() { - return defaultInstance; - } - - public IOSpecProto getDefaultInstanceForType() { - return defaultInstance; - } - - private final com.google.protobuf.UnknownFieldSet unknownFields; - @java.lang.Override - public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { - return this.unknownFields; - } - private IOSpecProto( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); - int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); - try { - boolean done = false; - while (!done) { - int tag = input.readTag(); - switch (tag) { - case 0: - done = true; - break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } - case 10: { - bitField0_ |= 0x00000001; - connectedVertexName_ = input.readBytes(); - break; - } - case 18: { - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder subBuilder = null; - if (((bitField0_ & 0x00000002) == 0x00000002)) { - subBuilder = ioDescriptor_.toBuilder(); - } - ioDescriptor_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.PARSER, extensionRegistry); - if (subBuilder != null) { - subBuilder.mergeFrom(ioDescriptor_); - ioDescriptor_ = subBuilder.buildPartial(); - } - bitField0_ |= 0x00000002; - break; - } - case 24: { - bitField0_ |= 0x00000004; - physicalEdgeCount_ = input.readInt32(); - break; - } - } - } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { - throw e.setUnfinishedMessage(this); - } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); - } finally { - this.unknownFields = unknownFields.build(); - makeExtensionsImmutable(); - } - } - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder.class); - } - - public static com.google.protobuf.Parser<IOSpecProto> PARSER = - new com.google.protobuf.AbstractParser<IOSpecProto>() { - public IOSpecProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new IOSpecProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser<IOSpecProto> getParserForType() { - return PARSER; - } - - private int bitField0_; - // optional string connected_vertex_name = 1; - public static final int CONNECTED_VERTEX_NAME_FIELD_NUMBER = 1; - private java.lang.Object connectedVertexName_; - /** - * <code>optional string connected_vertex_name = 1;</code> - */ - public boolean hasConnectedVertexName() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * <code>optional string connected_vertex_name = 1;</code> - */ - public java.lang.String getConnectedVertexName() { - java.lang.Object ref = connectedVertexName_; - if (ref instanceof java.lang.String) { - return (java.lang.String) ref; - } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; - java.lang.String s = bs.toStringUtf8(); - if (bs.isValidUtf8()) { - connectedVertexName_ = s; - } - return s; - } - } - /** - * <code>optional string connected_vertex_name = 1;</code> - */ - public com.google.protobuf.ByteString - getConnectedVertexNameBytes() { - java.lang.Object ref = connectedVertexName_; - if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - connectedVertexName_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - - // optional .EntityDescriptorProto io_descriptor = 2; - public static final int IO_DESCRIPTOR_FIELD_NUMBER = 2; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto ioDescriptor_; - /** - * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> - */ - public boolean hasIoDescriptor() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getIoDescriptor() { - return ioDescriptor_; - } - /** - * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getIoDescriptorOrBuilder() { - return ioDescriptor_; - } - - // optional int32 physical_edge_count = 3; - public static final int PHYSICAL_EDGE_COUNT_FIELD_NUMBER = 3; - private int physicalEdgeCount_; - /** - * <code>optional int32 physical_edge_count = 3;</code> - */ - public boolean hasPhysicalEdgeCount() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - /** - * <code>optional int32 physical_edge_count = 3;</code> - */ - public int getPhysicalEdgeCount() { - return physicalEdgeCount_; - } - - private void initFields() { - connectedVertexName_ = ""; - ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); - physicalEdgeCount_ = 0; - } - private byte memoizedIsInitialized = -1; - public final boolean isInitialized() { - byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; - - memoizedIsInitialized = 1; - return true; - } - - public void writeTo(com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getConnectedVertexNameBytes()); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeMessage(2, ioDescriptor_); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeInt32(3, physicalEdgeCount_); - } - getUnknownFields().writeTo(output); - } - - private int memoizedSerializedSize = -1; - public int getSerializedSize() { - int size = memoizedSerializedSize; - if (size != -1) return size; - - size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getConnectedVertexNameBytes()); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(2, ioDescriptor_); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream - .computeInt32Size(3, physicalEdgeCount_); - } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; - return size; - } - - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override - public boolean equals(final java.lang.Object obj) { - if (obj == this) { - return true; - } - if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto)) { - return super.equals(obj); - } - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto) obj; - - boolean result = true; - result = result && (hasConnectedVertexName() == other.hasConnectedVertexName()); - if (hasConnectedVertexName()) { - result = result && getConnectedVertexName() - .equals(other.getConnectedVertexName()); - } - result = result && (hasIoDescriptor() == other.hasIoDescriptor()); - if (hasIoDescriptor()) { - result = result && getIoDescriptor() - .equals(other.getIoDescriptor()); - } - result = result && (hasPhysicalEdgeCount() == other.hasPhysicalEdgeCount()); - if (hasPhysicalEdgeCount()) { - result = result && (getPhysicalEdgeCount() - == other.getPhysicalEdgeCount()); - } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; - } - - private int memoizedHashCode = 0; - @java.lang.Override - public int hashCode() { - if (memoizedHashCode != 0) { - return memoizedHashCode; - } - int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasConnectedVertexName()) { - hash = (37 * hash) + CONNECTED_VERTEX_NAME_FIELD_NUMBER; - hash = (53 * hash) + getConnectedVertexName().hashCode(); - } - if (hasIoDescriptor()) { - hash = (37 * hash) + IO_DESCRIPTOR_FIELD_NUMBER; - hash = (53 * hash) + getIoDescriptor().hashCode(); - } - if (hasPhysicalEdgeCount()) { - hash = (37 * hash) + PHYSICAL_EDGE_COUNT_FIELD_NUMBER; - hash = (53 * hash) + getPhysicalEdgeCount(); - } - hash = (29 * hash) + getUnknownFields().hashCode(); - memoizedHashCode = hash; - return hash; - } - - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom( - byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(java.io.InputStream input) - throws java.io.IOException { - return PARSER.parseFrom(input); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseDelimitedFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom( - com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return PARSER.parseFrom(input); - } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); - } - - public static Builder newBuilder() { return Builder.create(); } - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto prototype) { - return newBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { return newBuilder(this); } - - @java.lang.Override - protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - Builder builder = new Builder(parent); - return builder; - } - /** - * Protobuf type {@code IOSpecProto} - */ - public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder<Builder> - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder.class); - } - - // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.newBuilder() - private Builder() { - maybeForceBuilderInitialization(); - } - - private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - super(parent); - maybeForceBuilderInitialization(); - } - private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { - getIoDescriptorFieldBuilder(); - } - } - private static Builder create() { - return new Builder(); - } - - public Builder clear() { - super.clear(); - connectedVertexName_ = ""; - bitField0_ = (bitField0_ & ~0x00000001); - if (ioDescriptorBuilder_ == null) { - ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); - } else { - ioDescriptorBuilder_.clear(); - } - bitField0_ = (bitField0_ & ~0x00000002); - physicalEdgeCount_ = 0; - bitField0_ = (bitField0_ & ~0x00000004); - return this; - } - - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - - public com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_descriptor; - } - - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getDefaultInstanceForType() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.getDefaultInstance(); - } - - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto build() { - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException(result); - } - return result; - } - - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto buildPartial() { - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto(this); - int from_bitField0_ = bitField0_; - int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { - to_bitField0_ |= 0x00000001; - } - result.connectedVertexName_ = connectedVertexName_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { - to_bitField0_ |= 0x00000002; - } - if (ioDescriptorBuilder_ == null) { - result.ioDescriptor_ = ioDescriptor_; - } else { - result.ioDescriptor_ = ioDescriptorBuilder_.build(); - } - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { - to_bitField0_ |= 0x00000004; - } - result.physicalEdgeCount_ = physicalEdgeCount_; - result.bitField0_ = to_bitField0_; - onBuilt(); - return result; - } - - public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto) { - return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto other) { - if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.getDefaultInstance()) return this; - if (other.hasConnectedVertexName()) { - bitField0_ |= 0x00000001; - connectedVertexName_ = other.connectedVertexName_; - onChanged(); - } - if (other.hasIoDescriptor()) { - mergeIoDescriptor(other.getIoDescriptor()); - } - if (other.hasPhysicalEdgeCount()) { - setPhysicalEdgeCount(other.getPhysicalEdgeCount()); - } - this.mergeUnknownFields(other.getUnknownFields()); - return this; - } - - public final boolean isInitialized() { - return true; - } - - public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parsedMessage = null; - try { - parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto) e.getUnfinishedMessage(); - throw e; - } finally { - if (parsedMessage != null) { - mergeFrom(parsedMessage); - } - } - return this; - } - private int bitField0_; - - // optional string connected_vertex_name = 1; - private java.lang.Object connectedVertexName_ = ""; - /** - * <code>optional string connected_vertex_name = 1;</code> - */ - public boolean hasConnectedVertexName() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * <code>optional string connected_vertex_name = 1;</code> - */ - public java.lang.String getConnectedVertexName() { - java.lang.Object ref = connectedVertexName_; - if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - connectedVertexName_ = s; - return s; - } else { - return (java.lang.String) ref; - } - } - /** - * <code>optional string connected_vertex_name = 1;</code> - */ - public com.google.protobuf.ByteString - getConnectedVertexNameBytes() { - java.lang.Object ref = connectedVertexName_; - if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - connectedVertexName_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - /** - * <code>optional string connected_vertex_name = 1;</code> - */ - public Builder setConnectedVertexName( - java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000001; - connectedVertexName_ = value; - onChanged(); - return this; - } - /** - * <code>optional string connected_vertex_name = 1;</code> - */ - public Builder clearConnectedVertexName() { - bitField0_ = (bitField0_ & ~0x00000001); - connectedVertexName_ = getDefaultInstance().getConnectedVertexName(); - onChanged(); - return this; - } - /** - * <code>optional string connected_vertex_name = 1;</code> - */ - public Builder setConnectedVertexNameBytes( - com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000001; - connectedVertexName_ = value; - onChanged(); - return this; - } - - // optional .EntityDescriptorProto io_descriptor = 2; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> ioDescriptorBuilder_; - /** - * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> - */ - public boolean hasIoDescriptor() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getIoDescriptor() { - if (ioDescriptorBuilder_ == null) { - return ioDescriptor_; - } else { - return ioDescriptorBuilder_.getMessage(); - } - } - /** - * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> - */ - public Builder setIoDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) { - if (ioDescriptorBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - ioDescriptor_ = value; - onChanged(); - } else { - ioDescriptorBuilder_.setMessage(value); - } - bitField0_ |= 0x00000002; - return this; - } - /** - * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> - */ - public Builder setIoDescriptor( - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder builderForValue) { - if (ioDescriptorBuilder_ == null) { - ioDescriptor_ = builderForValue.build(); - onChanged(); - } else { - ioDescriptorBuilder_.setMessage(builderForValue.build()); - } - bitField0_ |= 0x00000002; - return this; - } - /** - * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> - */ - public Builder mergeIoDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) { - if (ioDescriptorBuilder_ == null) { - if (((bitField0_ & 0x00000002) == 0x00000002) && - ioDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) { - ioDescriptor_ = - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.newBuilder(ioDescriptor_).mergeFrom(value).buildPartial(); - } else { - ioDescriptor_ = value; - } - onChanged(); - } else { - ioDescriptorBuilder_.mergeFrom(value); - } - bitField0_ |= 0x00000002; - return this; - } - /** - * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> - */ - public Builder clearIoDescriptor() { - if (ioDescriptorBuilder_ == null) { - ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); - onChanged(); - } else { - ioDescriptorBuilder_.clear(); - } - bitField0_ = (bitField0_ & ~0x00000002); - return this; - } - /** - * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder getIoDescriptorBuilder() { - bitField0_ |= 0x00000002; - onChanged(); - return getIoDescriptorFieldBuilder().getBuilder(); - } - /** - * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getIoDescriptorOrBuilder() { - if (ioDescriptorBuilder_ != null) { - return ioDescriptorBuilder_.getMessageOrBuilder(); - } else { - return ioDescriptor_; - } - } - /** - * <code>optional .EntityDescriptorProto io_descriptor = 2;</code> - */ - private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> - getIoDescriptorFieldBuilder() { - if (ioDescriptorBuilder_ == null) { - ioDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder>( - ioDescriptor_, - getParentForChildren(), - isClean()); - ioDescriptor_ = null; - } - return ioDescriptorBuilder_; - } - - // optional int32 physical_edge_count = 3; - private int physicalEdgeCount_ ; - /** - * <code>optional int32 physical_edge
<TRUNCATED>