http://git-wip-us.apache.org/repos/asf/hbase/blob/b4a729ed/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RPCProtos.java ---------------------------------------------------------------------- diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RPCProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RPCProtos.java index 222f383..ead61ed 100644 --- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RPCProtos.java +++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/RPCProtos.java @@ -6,12 +6,18 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated; public final class RPCProtos { private RPCProtos() {} public static void registerAllExtensions( + com.google.protobuf.ExtensionRegistryLite registry) { + } + + public static void registerAllExtensions( com.google.protobuf.ExtensionRegistry registry) { + registerAllExtensions( + (com.google.protobuf.ExtensionRegistryLite) registry); } - public interface UserInformationOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface UserInformationOrBuilder extends + // @@protoc_insertion_point(interface_extends:hbase.pb.UserInformation) + com.google.protobuf.MessageOrBuilder { - // required string effective_user = 1; /** * <code>required string effective_user = 1;</code> */ @@ -26,7 +32,6 @@ public final class RPCProtos { com.google.protobuf.ByteString getEffectiveUserBytes(); - // optional string real_user = 2; /** * <code>optional string real_user = 2;</code> */ @@ -42,42 +47,35 @@ public final class RPCProtos { getRealUserBytes(); } /** - * Protobuf type {@code hbase.pb.UserInformation} - * * <pre> * User Information proto. Included in ConnectionHeader on connection setup * </pre> + * + * Protobuf type {@code hbase.pb.UserInformation} */ - public static final class UserInformation extends - com.google.protobuf.GeneratedMessage - implements UserInformationOrBuilder { + public static final class UserInformation extends + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:hbase.pb.UserInformation) + UserInformationOrBuilder { // Use UserInformation.newBuilder() to construct. - private UserInformation(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + private UserInformation(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); - } - private UserInformation(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final UserInformation defaultInstance; - public static UserInformation getDefaultInstance() { - return defaultInstance; } - - public UserInformation getDefaultInstanceForType() { - return defaultInstance; + private UserInformation() { + effectiveUser_ = ""; + realUser_ = ""; } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private UserInformation( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -97,13 +95,15 @@ public final class RPCProtos { break; } case 10: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; - effectiveUser_ = input.readBytes(); + effectiveUser_ = bs; break; } case 18: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; - realUser_ = input.readBytes(); + realUser_ = bs; break; } } @@ -112,7 +112,7 @@ public final class RPCProtos { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -123,32 +123,16 @@ public final class RPCProtos { return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_UserInformation_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_UserInformation_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.class, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder.class); } - public static com.google.protobuf.Parser<UserInformation> PARSER = - new com.google.protobuf.AbstractParser<UserInformation>() { - public UserInformation parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new UserInformation(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser<UserInformation> getParserForType() { - return PARSER; - } - private int bitField0_; - // required string effective_user = 1; public static final int EFFECTIVE_USER_FIELD_NUMBER = 1; - private java.lang.Object effectiveUser_; + private volatile java.lang.Object effectiveUser_; /** * <code>required string effective_user = 1;</code> */ @@ -189,9 +173,8 @@ public final class RPCProtos { } } - // optional string real_user = 2; public static final int REAL_USER_FIELD_NUMBER = 2; - private java.lang.Object realUser_; + private volatile java.lang.Object realUser_; /** * <code>optional string real_user = 2;</code> */ @@ -232,14 +215,11 @@ public final class RPCProtos { } } - private void initFields() { - effectiveUser_ = ""; - realUser_ = ""; - } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; if (!hasEffectiveUser()) { memoizedIsInitialized = 0; @@ -251,43 +231,33 @@ public final class RPCProtos { public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getEffectiveUserBytes()); + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, effectiveUser_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, getRealUserBytes()); + com.google.protobuf.GeneratedMessageV3.writeString(output, 2, realUser_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getEffectiveUserBytes()); + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, effectiveUser_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, getRealUserBytes()); + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(2, realUser_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; @@ -308,12 +278,10 @@ public final class RPCProtos { result = result && getRealUser() .equals(other.getRealUser()); } - result = result && - getUnknownFields().equals(other.getUnknownFields()); + result = result && unknownFields.equals(other.unknownFields); return result; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { @@ -329,7 +297,7 @@ public final class RPCProtos { hash = (37 * hash) + REAL_USER_FIELD_NUMBER; hash = (53 * hash) + getRealUser().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } @@ -357,65 +325,77 @@ public final class RPCProtos { } public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** - * Protobuf type {@code hbase.pb.UserInformation} - * * <pre> * User Information proto. Included in ConnectionHeader on connection setup * </pre> + * + * Protobuf type {@code hbase.pb.UserInformation} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder<Builder> - implements org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements + // @@protoc_insertion_point(builder_implements:hbase.pb.UserInformation) + org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_UserInformation_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_UserInformation_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -428,18 +408,15 @@ public final class RPCProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - public Builder clear() { super.clear(); effectiveUser_ = ""; @@ -449,10 +426,6 @@ public final class RPCProtos { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_UserInformation_descriptor; @@ -487,6 +460,32 @@ public final class RPCProtos { return result; } + public Builder clone() { + return (Builder) super.clone(); + } + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + Object value) { + return (Builder) super.setField(field, value); + } + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return (Builder) super.clearField(field); + } + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return (Builder) super.clearOneof(oneof); + } + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, Object value) { + return (Builder) super.setRepeatedField(field, index, value); + } + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + Object value) { + return (Builder) super.addRepeatedField(field, value); + } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation)other); @@ -508,13 +507,13 @@ public final class RPCProtos { realUser_ = other.realUser_; onChanged(); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } public final boolean isInitialized() { if (!hasEffectiveUser()) { - return false; } return true; @@ -529,7 +528,7 @@ public final class RPCProtos { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -539,7 +538,6 @@ public final class RPCProtos { } private int bitField0_; - // required string effective_user = 1; private java.lang.Object effectiveUser_ = ""; /** * <code>required string effective_user = 1;</code> @@ -553,9 +551,12 @@ public final class RPCProtos { public java.lang.String getEffectiveUser() { java.lang.Object ref = effectiveUser_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - effectiveUser_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + effectiveUser_ = s; + } return s; } else { return (java.lang.String) ref; @@ -613,7 +614,6 @@ public final class RPCProtos { return this; } - // optional string real_user = 2; private java.lang.Object realUser_ = ""; /** * <code>optional string real_user = 2;</code> @@ -627,9 +627,12 @@ public final class RPCProtos { public java.lang.String getRealUser() { java.lang.Object ref = realUser_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - realUser_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + realUser_ = s; + } return s; } else { return (java.lang.String) ref; @@ -686,22 +689,59 @@ public final class RPCProtos { onChanged(); return this; } + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:hbase.pb.UserInformation) } + // @@protoc_insertion_point(class_scope:hbase.pb.UserInformation) + private static final org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation DEFAULT_INSTANCE; static { - defaultInstance = new UserInformation(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation(); + } + + public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser<UserInformation> + PARSER = new com.google.protobuf.AbstractParser<UserInformation>() { + public UserInformation parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new UserInformation(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser<UserInformation> parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser<UserInformation> getParserForType() { + return PARSER; + } + + public org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:hbase.pb.UserInformation) } - public interface ConnectionHeaderOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface ConnectionHeaderOrBuilder extends + // @@protoc_insertion_point(interface_extends:hbase.pb.ConnectionHeader) + com.google.protobuf.MessageOrBuilder { - // optional .hbase.pb.UserInformation user_info = 1; /** * <code>optional .hbase.pb.UserInformation user_info = 1;</code> */ @@ -715,7 +755,6 @@ public final class RPCProtos { */ org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder getUserInfoOrBuilder(); - // optional string service_name = 2; /** * <code>optional string service_name = 2;</code> */ @@ -730,67 +769,64 @@ public final class RPCProtos { com.google.protobuf.ByteString getServiceNameBytes(); - // optional string cell_block_codec_class = 3; /** - * <code>optional string cell_block_codec_class = 3;</code> - * * <pre> * Cell block codec we will use sending over optional cell blocks. Server throws exception * if cannot deal. Null means no codec'ing going on so we are pb all the time (SLOW!!!) * </pre> + * + * <code>optional string cell_block_codec_class = 3;</code> */ boolean hasCellBlockCodecClass(); /** - * <code>optional string cell_block_codec_class = 3;</code> - * * <pre> * Cell block codec we will use sending over optional cell blocks. Server throws exception * if cannot deal. Null means no codec'ing going on so we are pb all the time (SLOW!!!) * </pre> + * + * <code>optional string cell_block_codec_class = 3;</code> */ java.lang.String getCellBlockCodecClass(); /** - * <code>optional string cell_block_codec_class = 3;</code> - * * <pre> * Cell block codec we will use sending over optional cell blocks. Server throws exception * if cannot deal. Null means no codec'ing going on so we are pb all the time (SLOW!!!) * </pre> + * + * <code>optional string cell_block_codec_class = 3;</code> */ com.google.protobuf.ByteString getCellBlockCodecClassBytes(); - // optional string cell_block_compressor_class = 4; /** - * <code>optional string cell_block_compressor_class = 4;</code> - * * <pre> * Compressor we will use if cell block is compressed. Server will throw exception if not supported. * Class must implement hadoop's CompressionCodec Interface. Can't compress if no codec. * </pre> + * + * <code>optional string cell_block_compressor_class = 4;</code> */ boolean hasCellBlockCompressorClass(); /** - * <code>optional string cell_block_compressor_class = 4;</code> - * * <pre> * Compressor we will use if cell block is compressed. Server will throw exception if not supported. * Class must implement hadoop's CompressionCodec Interface. Can't compress if no codec. * </pre> + * + * <code>optional string cell_block_compressor_class = 4;</code> */ java.lang.String getCellBlockCompressorClass(); /** - * <code>optional string cell_block_compressor_class = 4;</code> - * * <pre> * Compressor we will use if cell block is compressed. Server will throw exception if not supported. * Class must implement hadoop's CompressionCodec Interface. Can't compress if no codec. * </pre> + * + * <code>optional string cell_block_compressor_class = 4;</code> */ com.google.protobuf.ByteString getCellBlockCompressorClassBytes(); - // optional .hbase.pb.VersionInfo version_info = 5; /** * <code>optional .hbase.pb.VersionInfo version_info = 5;</code> */ @@ -805,42 +841,36 @@ public final class RPCProtos { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfoOrBuilder getVersionInfoOrBuilder(); } /** - * Protobuf type {@code hbase.pb.ConnectionHeader} - * * <pre> * This is sent on connection setup after the connection preamble is sent. * </pre> + * + * Protobuf type {@code hbase.pb.ConnectionHeader} */ - public static final class ConnectionHeader extends - com.google.protobuf.GeneratedMessage - implements ConnectionHeaderOrBuilder { + public static final class ConnectionHeader extends + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:hbase.pb.ConnectionHeader) + ConnectionHeaderOrBuilder { // Use ConnectionHeader.newBuilder() to construct. - private ConnectionHeader(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + private ConnectionHeader(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private ConnectionHeader(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final ConnectionHeader defaultInstance; - public static ConnectionHeader getDefaultInstance() { - return defaultInstance; - } - - public ConnectionHeader getDefaultInstanceForType() { - return defaultInstance; + private ConnectionHeader() { + serviceName_ = ""; + cellBlockCodecClass_ = ""; + cellBlockCompressorClass_ = ""; } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private ConnectionHeader( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -873,18 +903,21 @@ public final class RPCProtos { break; } case 18: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; - serviceName_ = input.readBytes(); + serviceName_ = bs; break; } case 26: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000004; - cellBlockCodecClass_ = input.readBytes(); + cellBlockCodecClass_ = bs; break; } case 34: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000008; - cellBlockCompressorClass_ = input.readBytes(); + cellBlockCompressorClass_ = bs; break; } case 42: { @@ -906,7 +939,7 @@ public final class RPCProtos { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -917,30 +950,14 @@ public final class RPCProtos { return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_ConnectionHeader_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_ConnectionHeader_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader.class, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader.Builder.class); } - public static com.google.protobuf.Parser<ConnectionHeader> PARSER = - new com.google.protobuf.AbstractParser<ConnectionHeader>() { - public ConnectionHeader parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new ConnectionHeader(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser<ConnectionHeader> getParserForType() { - return PARSER; - } - private int bitField0_; - // optional .hbase.pb.UserInformation user_info = 1; public static final int USER_INFO_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation userInfo_; /** @@ -953,18 +970,17 @@ public final class RPCProtos { * <code>optional .hbase.pb.UserInformation user_info = 1;</code> */ public org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation getUserInfo() { - return userInfo_; + return userInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance() : userInfo_; } /** * <code>optional .hbase.pb.UserInformation user_info = 1;</code> */ public org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder getUserInfoOrBuilder() { - return userInfo_; + return userInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance() : userInfo_; } - // optional string service_name = 2; public static final int SERVICE_NAME_FIELD_NUMBER = 2; - private java.lang.Object serviceName_; + private volatile java.lang.Object serviceName_; /** * <code>optional string service_name = 2;</code> */ @@ -1005,27 +1021,26 @@ public final class RPCProtos { } } - // optional string cell_block_codec_class = 3; public static final int CELL_BLOCK_CODEC_CLASS_FIELD_NUMBER = 3; - private java.lang.Object cellBlockCodecClass_; + private volatile java.lang.Object cellBlockCodecClass_; /** - * <code>optional string cell_block_codec_class = 3;</code> - * * <pre> * Cell block codec we will use sending over optional cell blocks. Server throws exception * if cannot deal. Null means no codec'ing going on so we are pb all the time (SLOW!!!) * </pre> + * + * <code>optional string cell_block_codec_class = 3;</code> */ public boolean hasCellBlockCodecClass() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** - * <code>optional string cell_block_codec_class = 3;</code> - * * <pre> * Cell block codec we will use sending over optional cell blocks. Server throws exception * if cannot deal. Null means no codec'ing going on so we are pb all the time (SLOW!!!) * </pre> + * + * <code>optional string cell_block_codec_class = 3;</code> */ public java.lang.String getCellBlockCodecClass() { java.lang.Object ref = cellBlockCodecClass_; @@ -1042,12 +1057,12 @@ public final class RPCProtos { } } /** - * <code>optional string cell_block_codec_class = 3;</code> - * * <pre> * Cell block codec we will use sending over optional cell blocks. Server throws exception * if cannot deal. Null means no codec'ing going on so we are pb all the time (SLOW!!!) * </pre> + * + * <code>optional string cell_block_codec_class = 3;</code> */ public com.google.protobuf.ByteString getCellBlockCodecClassBytes() { @@ -1063,27 +1078,26 @@ public final class RPCProtos { } } - // optional string cell_block_compressor_class = 4; public static final int CELL_BLOCK_COMPRESSOR_CLASS_FIELD_NUMBER = 4; - private java.lang.Object cellBlockCompressorClass_; + private volatile java.lang.Object cellBlockCompressorClass_; /** - * <code>optional string cell_block_compressor_class = 4;</code> - * * <pre> * Compressor we will use if cell block is compressed. Server will throw exception if not supported. * Class must implement hadoop's CompressionCodec Interface. Can't compress if no codec. * </pre> + * + * <code>optional string cell_block_compressor_class = 4;</code> */ public boolean hasCellBlockCompressorClass() { return ((bitField0_ & 0x00000008) == 0x00000008); } /** - * <code>optional string cell_block_compressor_class = 4;</code> - * * <pre> * Compressor we will use if cell block is compressed. Server will throw exception if not supported. * Class must implement hadoop's CompressionCodec Interface. Can't compress if no codec. * </pre> + * + * <code>optional string cell_block_compressor_class = 4;</code> */ public java.lang.String getCellBlockCompressorClass() { java.lang.Object ref = cellBlockCompressorClass_; @@ -1100,12 +1114,12 @@ public final class RPCProtos { } } /** - * <code>optional string cell_block_compressor_class = 4;</code> - * * <pre> * Compressor we will use if cell block is compressed. Server will throw exception if not supported. * Class must implement hadoop's CompressionCodec Interface. Can't compress if no codec. * </pre> + * + * <code>optional string cell_block_compressor_class = 4;</code> */ public com.google.protobuf.ByteString getCellBlockCompressorClassBytes() { @@ -1121,7 +1135,6 @@ public final class RPCProtos { } } - // optional .hbase.pb.VersionInfo version_info = 5; public static final int VERSION_INFO_FIELD_NUMBER = 5; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo versionInfo_; /** @@ -1134,26 +1147,20 @@ public final class RPCProtos { * <code>optional .hbase.pb.VersionInfo version_info = 5;</code> */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo getVersionInfo() { - return versionInfo_; + return versionInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo.getDefaultInstance() : versionInfo_; } /** * <code>optional .hbase.pb.VersionInfo version_info = 5;</code> */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfoOrBuilder getVersionInfoOrBuilder() { - return versionInfo_; + return versionInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo.getDefaultInstance() : versionInfo_; } - private void initFields() { - userInfo_ = org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance(); - serviceName_ = ""; - cellBlockCodecClass_ = ""; - cellBlockCompressorClass_ = ""; - versionInfo_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo.getDefaultInstance(); - } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; if (hasUserInfo()) { if (!getUserInfo().isInitialized()) { @@ -1173,64 +1180,53 @@ public final class RPCProtos { public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeMessage(1, userInfo_); + output.writeMessage(1, getUserInfo()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, getServiceNameBytes()); + com.google.protobuf.GeneratedMessageV3.writeString(output, 2, serviceName_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeBytes(3, getCellBlockCodecClassBytes()); + com.google.protobuf.GeneratedMessageV3.writeString(output, 3, cellBlockCodecClass_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { - output.writeBytes(4, getCellBlockCompressorClassBytes()); + com.google.protobuf.GeneratedMessageV3.writeString(output, 4, cellBlockCompressorClass_); } if (((bitField0_ & 0x00000010) == 0x00000010)) { - output.writeMessage(5, versionInfo_); + output.writeMessage(5, getVersionInfo()); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(1, userInfo_); + .computeMessageSize(1, getUserInfo()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, getServiceNameBytes()); + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(2, serviceName_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(3, getCellBlockCodecClassBytes()); + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(3, cellBlockCodecClass_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(4, getCellBlockCompressorClassBytes()); + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(4, cellBlockCompressorClass_); } if (((bitField0_ & 0x00000010) == 0x00000010)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(5, versionInfo_); + .computeMessageSize(5, getVersionInfo()); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; @@ -1266,12 +1262,10 @@ public final class RPCProtos { result = result && getVersionInfo() .equals(other.getVersionInfo()); } - result = result && - getUnknownFields().equals(other.getUnknownFields()); + result = result && unknownFields.equals(other.unknownFields); return result; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { @@ -1299,7 +1293,7 @@ public final class RPCProtos { hash = (37 * hash) + VERSION_INFO_FIELD_NUMBER; hash = (53 * hash) + getVersionInfo().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } @@ -1327,65 +1321,77 @@ public final class RPCProtos { } public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** - * Protobuf type {@code hbase.pb.ConnectionHeader} - * * <pre> * This is sent on connection setup after the connection preamble is sent. * </pre> + * + * Protobuf type {@code hbase.pb.ConnectionHeader} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder<Builder> - implements org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeaderOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements + // @@protoc_insertion_point(builder_implements:hbase.pb.ConnectionHeader) + org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeaderOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_ConnectionHeader_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_ConnectionHeader_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -1398,24 +1404,21 @@ public final class RPCProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getUserInfoFieldBuilder(); getVersionInfoFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - public Builder clear() { super.clear(); if (userInfoBuilder_ == null) { - userInfo_ = org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance(); + userInfo_ = null; } else { userInfoBuilder_.clear(); } @@ -1427,7 +1430,7 @@ public final class RPCProtos { cellBlockCompressorClass_ = ""; bitField0_ = (bitField0_ & ~0x00000008); if (versionInfoBuilder_ == null) { - versionInfo_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo.getDefaultInstance(); + versionInfo_ = null; } else { versionInfoBuilder_.clear(); } @@ -1435,10 +1438,6 @@ public final class RPCProtos { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_ConnectionHeader_descriptor; @@ -1493,6 +1492,32 @@ public final class RPCProtos { return result; } + public Builder clone() { + return (Builder) super.clone(); + } + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + Object value) { + return (Builder) super.setField(field, value); + } + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return (Builder) super.clearField(field); + } + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return (Builder) super.clearOneof(oneof); + } + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, Object value) { + return (Builder) super.setRepeatedField(field, index, value); + } + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + Object value) { + return (Builder) super.addRepeatedField(field, value); + } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader)other); @@ -1525,20 +1550,19 @@ public final class RPCProtos { if (other.hasVersionInfo()) { mergeVersionInfo(other.getVersionInfo()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } public final boolean isInitialized() { if (hasUserInfo()) { if (!getUserInfo().isInitialized()) { - return false; } } if (hasVersionInfo()) { if (!getVersionInfo().isInitialized()) { - return false; } } @@ -1554,7 +1578,7 @@ public final class RPCProtos { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -1564,9 +1588,8 @@ public final class RPCProtos { } private int bitField0_; - // optional .hbase.pb.UserInformation user_info = 1; - private org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation userInfo_ = org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation userInfo_ = null; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder> userInfoBuilder_; /** * <code>optional .hbase.pb.UserInformation user_info = 1;</code> @@ -1579,7 +1602,7 @@ public final class RPCProtos { */ public org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation getUserInfo() { if (userInfoBuilder_ == null) { - return userInfo_; + return userInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance() : userInfo_; } else { return userInfoBuilder_.getMessage(); } @@ -1620,6 +1643,7 @@ public final class RPCProtos { public Builder mergeUserInfo(org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation value) { if (userInfoBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && + userInfo_ != null && userInfo_ != org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance()) { userInfo_ = org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.newBuilder(userInfo_).mergeFrom(value).buildPartial(); @@ -1638,7 +1662,7 @@ public final class RPCProtos { */ public Builder clearUserInfo() { if (userInfoBuilder_ == null) { - userInfo_ = org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance(); + userInfo_ = null; onChanged(); } else { userInfoBuilder_.clear(); @@ -1661,19 +1685,20 @@ public final class RPCProtos { if (userInfoBuilder_ != null) { return userInfoBuilder_.getMessageOrBuilder(); } else { - return userInfo_; + return userInfo_ == null ? + org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance() : userInfo_; } } /** * <code>optional .hbase.pb.UserInformation user_info = 1;</code> */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder> getUserInfoFieldBuilder() { if (userInfoBuilder_ == null) { - userInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder< + userInfoBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformation.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.UserInformationOrBuilder>( - userInfo_, + getUserInfo(), getParentForChildren(), isClean()); userInfo_ = null; @@ -1681,7 +1706,6 @@ public final class RPCProtos { return userInfoBuilder_; } - // optional string service_name = 2; private java.lang.Object serviceName_ = ""; /** * <code>optional string service_name = 2;</code> @@ -1695,9 +1719,12 @@ public final class RPCProtos { public java.lang.String getServiceName() { java.lang.Object ref = serviceName_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - serviceName_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + serviceName_ = s; + } return s; } else { return (java.lang.String) ref; @@ -1755,45 +1782,47 @@ public final class RPCProtos { return this; } - // optional string cell_block_codec_class = 3; private java.lang.Object cellBlockCodecClass_ = ""; /** - * <code>optional string cell_block_codec_class = 3;</code> - * * <pre> * Cell block codec we will use sending over optional cell blocks. Server throws exception * if cannot deal. Null means no codec'ing going on so we are pb all the time (SLOW!!!) * </pre> + * + * <code>optional string cell_block_codec_class = 3;</code> */ public boolean hasCellBlockCodecClass() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** - * <code>optional string cell_block_codec_class = 3;</code> - * * <pre> * Cell block codec we will use sending over optional cell blocks. Server throws exception * if cannot deal. Null means no codec'ing going on so we are pb all the time (SLOW!!!) * </pre> + * + * <code>optional string cell_block_codec_class = 3;</code> */ public java.lang.String getCellBlockCodecClass() { java.lang.Object ref = cellBlockCodecClass_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - cellBlockCodecClass_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + cellBlockCodecClass_ = s; + } return s; } else { return (java.lang.String) ref; } } /** - * <code>optional string cell_block_codec_class = 3;</code> - * * <pre> * Cell block codec we will use sending over optional cell blocks. Server throws exception * if cannot deal. Null means no codec'ing going on so we are pb all the time (SLOW!!!) * </pre> + * + * <code>optional string cell_block_codec_class = 3;</code> */ public com.google.protobuf.ByteString getCellBlockCodecClassBytes() { @@ -1809,12 +1838,12 @@ public final class RPCProtos { } } /** - * <code>optional string cell_block_codec_class = 3;</code> - * * <pre> * Cell block codec we will use sending over optional cell blocks. Server throws exception * if cannot deal. Null means no codec'ing going on so we are pb all the time (SLOW!!!) * </pre> + * + * <code>optional string cell_block_codec_class = 3;</code> */ public Builder setCellBlockCodecClass( java.lang.String value) { @@ -1827,12 +1856,12 @@ public final class RPCProtos { return this; } /** - * <code>optional string cell_block_codec_class = 3;</code> - * * <pre> * Cell block codec we will use sending over optional cell blocks. Server throws exception * if cannot deal. Null means no codec'ing going on so we are pb all the time (SLOW!!!) * </pre> + * + * <code>optional string cell_block_codec_class = 3;</code> */ public Builder clearCellBlockCodecClass() { bitField0_ = (bitField0_ & ~0x00000004); @@ -1841,12 +1870,12 @@ public final class RPCProtos { return this; } /** - * <code>optional string cell_block_codec_class = 3;</code> - * * <pre> * Cell block codec we will use sending over optional cell blocks. Server throws exception * if cannot deal. Null means no codec'ing going on so we are pb all the time (SLOW!!!) * </pre> + * + * <code>optional string cell_block_codec_class = 3;</code> */ public Builder setCellBlockCodecClassBytes( com.google.protobuf.ByteString value) { @@ -1859,45 +1888,47 @@ public final class RPCProtos { return this; } - // optional string cell_block_compressor_class = 4; private java.lang.Object cellBlockCompressorClass_ = ""; /** - * <code>optional string cell_block_compressor_class = 4;</code> - * * <pre> * Compressor we will use if cell block is compressed. Server will throw exception if not supported. * Class must implement hadoop's CompressionCodec Interface. Can't compress if no codec. * </pre> + * + * <code>optional string cell_block_compressor_class = 4;</code> */ public boolean hasCellBlockCompressorClass() { return ((bitField0_ & 0x00000008) == 0x00000008); } /** - * <code>optional string cell_block_compressor_class = 4;</code> - * * <pre> * Compressor we will use if cell block is compressed. Server will throw exception if not supported. * Class must implement hadoop's CompressionCodec Interface. Can't compress if no codec. * </pre> + * + * <code>optional string cell_block_compressor_class = 4;</code> */ public java.lang.String getCellBlockCompressorClass() { java.lang.Object ref = cellBlockCompressorClass_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - cellBlockCompressorClass_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + cellBlockCompressorClass_ = s; + } return s; } else { return (java.lang.String) ref; } } /** - * <code>optional string cell_block_compressor_class = 4;</code> - * * <pre> * Compressor we will use if cell block is compressed. Server will throw exception if not supported. * Class must implement hadoop's CompressionCodec Interface. Can't compress if no codec. * </pre> + * + * <code>optional string cell_block_compressor_class = 4;</code> */ public com.google.protobuf.ByteString getCellBlockCompressorClassBytes() { @@ -1913,12 +1944,12 @@ public final class RPCProtos { } } /** - * <code>optional string cell_block_compressor_class = 4;</code> - * * <pre> * Compressor we will use if cell block is compressed. Server will throw exception if not supported. * Class must implement hadoop's CompressionCodec Interface. Can't compress if no codec. * </pre> + * + * <code>optional string cell_block_compressor_class = 4;</code> */ public Builder setCellBlockCompressorClass( java.lang.String value) { @@ -1931,12 +1962,12 @@ public final class RPCProtos { return this; } /** - * <code>optional string cell_block_compressor_class = 4;</code> - * * <pre> * Compressor we will use if cell block is compressed. Server will throw exception if not supported. * Class must implement hadoop's CompressionCodec Interface. Can't compress if no codec. * </pre> + * + * <code>optional string cell_block_compressor_class = 4;</code> */ public Builder clearCellBlockCompressorClass() { bitField0_ = (bitField0_ & ~0x00000008); @@ -1945,12 +1976,12 @@ public final class RPCProtos { return this; } /** - * <code>optional string cell_block_compressor_class = 4;</code> - * * <pre> * Compressor we will use if cell block is compressed. Server will throw exception if not supported. * Class must implement hadoop's CompressionCodec Interface. Can't compress if no codec. * </pre> + * + * <code>optional string cell_block_compressor_class = 4;</code> */ public Builder setCellBlockCompressorClassBytes( com.google.protobuf.ByteString value) { @@ -1963,9 +1994,8 @@ public final class RPCProtos { return this; } - // optional .hbase.pb.VersionInfo version_info = 5; - private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo versionInfo_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo versionInfo_ = null; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfoOrBuilder> versionInfoBuilder_; /** * <code>optional .hbase.pb.VersionInfo version_info = 5;</code> @@ -1978,7 +2008,7 @@ public final class RPCProtos { */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo getVersionInfo() { if (versionInfoBuilder_ == null) { - return versionInfo_; + return versionInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo.getDefaultInstance() : versionInfo_; } else { return versionInfoBuilder_.getMessage(); } @@ -2019,6 +2049,7 @@ public final class RPCProtos { public Builder mergeVersionInfo(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo value) { if (versionInfoBuilder_ == null) { if (((bitField0_ & 0x00000010) == 0x00000010) && + versionInfo_ != null && versionInfo_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo.getDefaultInstance()) { versionInfo_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo.newBuilder(versionInfo_).mergeFrom(value).buildPartial(); @@ -2037,7 +2068,7 @@ public final class RPCProtos { */ public Builder clearVersionInfo() { if (versionInfoBuilder_ == null) { - versionInfo_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo.getDefaultInstance(); + versionInfo_ = null; onChanged(); } else { versionInfoBuilder_.clear(); @@ -2060,95 +2091,125 @@ public final class RPCProtos { if (versionInfoBuilder_ != null) { return versionInfoBuilder_.getMessageOrBuilder(); } else { - return versionInfo_; + return versionInfo_ == null ? + org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo.getDefaultInstance() : versionInfo_; } } /** * <code>optional .hbase.pb.VersionInfo version_info = 5;</code> */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfoOrBuilder> getVersionInfoFieldBuilder() { if (versionInfoBuilder_ == null) { - versionInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder< + versionInfoBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.VersionInfoOrBuilder>( - versionInfo_, + getVersionInfo(), getParentForChildren(), isClean()); versionInfo_ = null; } return versionInfoBuilder_; } + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:hbase.pb.ConnectionHeader) } + // @@protoc_insertion_point(class_scope:hbase.pb.ConnectionHeader) + private static final org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader DEFAULT_INSTANCE; static { - defaultInstance = new ConnectionHeader(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader(); } - // @@protoc_insertion_point(class_scope:hbase.pb.ConnectionHeader) - } + public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader getDefaultInstance() { + return DEFAULT_INSTANCE; + } - public interface CellBlockMetaOrBuilder - extends com.google.protobuf.MessageOrBuilder { + @java.lang.Deprecated public static final com.google.protobuf.Parser<ConnectionHeader> + PARSER = new com.google.protobuf.AbstractParser<ConnectionHeader>() { + public ConnectionHeader parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new ConnectionHeader(input, extensionRegistry); + } + }; - // optional uint32 length = 1; - /** - * <code>optional uint32 length = 1;</code> - * - * <pre> - * Length of the following cell block. Could calculate it but convenient having it too hand. + public static com.google.protobuf.Parser<ConnectionHeader> parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser<ConnectionHeader> getParserForType() { + return PARSER; + } + + public org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.ConnectionHeader getDefaultInstanceForType() { + return DEFAULT_INSTANCE; + } + + } + + public interface CellBlockMetaOrBuilder extends + // @@protoc_insertion_point(interface_extends:hbase.pb.CellBlockMeta) + com.google.protobuf.MessageOrBuilder { + + /** + * <pre> + * Length of the following cell block. Could calculate it but convenient having it too hand. * </pre> + * + * <code>optional uint32 length = 1;</code> */ boolean hasLength(); /** - * <code>optional uint32 length = 1;</code> - * * <pre> * Length of the following cell block. Could calculate it but convenient having it too hand. * </pre> + * + * <code>optional uint32 length = 1;</code> */ int getLength(); } /** - * Protobuf type {@code hbase.pb.CellBlockMeta} - * * <pre> * Optional Cell block Message. Included in client RequestHeader * </pre> + * + * Protobuf type {@code hbase.pb.CellBlockMeta} */ - public static final class CellBlockMeta extends - com.google.protobuf.GeneratedMessage - implements CellBlockMetaOrBuilder { + public static final class CellBlockMeta extends + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:hbase.pb.CellBlockMeta) + CellBlockMetaOrBuilder { // Use CellBlockMeta.newBuilder() to construct. - private CellBlockMeta(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + private CellBlockMeta(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); - } - private CellBlockMeta(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final CellBlockMeta defaultInstance; - public static CellBlockMeta getDefaultInstance() { - return defaultInstance; } - - public CellBlockMeta getDefaultInstanceForType() { - return defaultInstance; + private CellBlockMeta() { + length_ = 0; } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private CellBlockMeta( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -2178,7 +2239,7 @@ public final class RPCProtos { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -2189,60 +2250,42 @@ public final class RPCProtos { return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_CellBlockMeta_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_CellBlockMeta_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta.class, org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta.Builder.class); } - public static com.google.protobuf.Parser<CellBlockMeta> PARSER = - new com.google.protobuf.AbstractParser<CellBlockMeta>() { - public CellBlockMeta parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new CellBlockMeta(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser<CellBlockMeta> getParserForType() { - return PARSER; - } - private int bitField0_; - // optional uint32 length = 1; public static final int LENGTH_FIELD_NUMBER = 1; private int length_; /** - * <code>optional uint32 length = 1;</code> - * * <pre> * Length of the following cell block. Could calculate it but convenient having it too hand. * </pre> + * + * <code>optional uint32 length = 1;</code> */ public boolean hasLength() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** - * <code>optional uint32 length = 1;</code> - * * <pre> * Length of the following cell block. Could calculate it but convenient having it too hand. * </pre> + * + * <code>optional uint32 length = 1;</code> */ public int getLength() { return length_; } - private void initFields() { - length_ = 0; - } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; @@ -2250,16 +2293,14 @@ public final class RPCProtos { public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeUInt32(1, length_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; @@ -2267,19 +2308,13 @@ public final class RPCProtos { size += com.google.protobuf.CodedOutputStream .computeUInt32Size(1, length_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; @@ -2295,12 +2330,10 @@ public final class RPCProtos { result = result && (getLength() == other.getLength()); } - result = result && - getUnknownFields().equals(other.getUnknownFields()); + result = result && unknownFields.equals(other.unknownFields); return result; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { @@ -2312,7 +2345,7 @@ public final class RPCProtos { hash = (37 * hash) + LENGTH_FIELD_NUMBER; hash = (53 * hash) + getLength(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } @@ -2340,65 +2373,77 @@ public final class RPCProtos { } public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** - * Protobuf type {@code hbase.pb.CellBlockMeta} - * * <pre> * Optional Cell block Message. Included in client RequestHeader * </pre> + * + * Protobuf type {@code hbase.pb.CellBlockMeta} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder<Builder> - implements org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements + // @@protoc_insertion_point(builder_implements:hbase.pb.CellBlockMeta) + org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMetaOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_CellBlockMeta_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_CellBlockMeta_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -2411,18 +2456,15 @@ public final class RPCProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - public Builder clear() { super.clear(); length_ = 0; @@ -2430,10 +2472,6 @@ public final class RPCProtos { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.internal_static_hbase_pb_CellBlockMeta_descriptor; @@ -2464,6 +2502,32 @@ public final class RPCProtos { return result; } + public Builder clone() { + return (Builder) super.clone(); + } + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + Object value) { + return (Builder) super.setField(field, value); + } + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return (Builder) super.clearField(field); + } + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return (Builder) super.clearOneof(oneof); + } + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, Object value) { + return (Builder) super.setRepeatedField(field, index, value); + } + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + Object value) { + return (Builder) super.addRepeatedField(field, value); + } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta)other); @@ -2478,7 +2542,8 @@ public final class RPCProtos { if (other.hasLength()) { setLength(other.getLength()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } @@ -2495,7 +2560,7 @@ public final class RPCProtos { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -2505,34 +2570,33 @@ public final class RPCProtos { } private int bitField0_; - // optional uint32 length = 1; private int length_ ; /** - * <code>optional uint32 length = 1;</code> - * * <pre> * Length of the following cell block. Could calculate it but convenient having it too hand. * </pre> + * + * <code>optional uint32 length = 1;</code> */ public boolean hasLength() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** - * <code>optional uint32 length = 1;</code> - * * <pre> * Length of the following cell block. Could calculate it but convenient having it too hand. * </pre> + * + * <code>optional uint32 length = 1;</code> */ public int getLength() { return length_; } /** - * <code>optional uint32 length = 1;</code> - * * <pre> * Length of the following cell block. Could calculate it but convenient having it too hand. * </pre> + * + * <code>optional uint32 length = 1;</code> */ public Builder setLength(int value) { bitField0_ |= 0x00000001; @@ -2541,11 +2605,11 @@ public final class RPCProtos { return this; } /** - * <code>optional uint32 length = 1;</code> - * * <pre> * Length of the following cell block. Could calculate it but convenient having it too hand. * </pre> + * + * <code>optional uint32 length = 1;</code> */ public Builder clearLength() { bitField0_ = (bitField0_ & ~0x00000001); @@ -2553,106 +2617,140 @@ public final class RPCProtos { onChanged(); return this; } + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:hbase.pb.CellBlockMeta) } + // @@protoc_insertion_point(class_scope:hbase.pb.CellBlockMeta) + private static final org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta DEFAULT_INSTANCE; static { - defaultInstance = new CellBlockMeta(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta(); + } + + public static org.apache.hadoop.hbase.shaded.protobuf.generated.RPCProtos.CellBlockMeta getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser<CellBlockMeta> + PARSER = new com.
<TRUNCATED>