http://git-wip-us.apache.org/repos/asf/hbase/blob/b4a729ed/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FilterProtos.java ---------------------------------------------------------------------- diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FilterProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FilterProtos.java index b63acb2..7204813 100644 --- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FilterProtos.java +++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FilterProtos.java @@ -6,12 +6,18 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated; public final class FilterProtos { private FilterProtos() {} public static void registerAllExtensions( + com.google.protobuf.ExtensionRegistryLite registry) { + } + + public static void registerAllExtensions( com.google.protobuf.ExtensionRegistry registry) { + registerAllExtensions( + (com.google.protobuf.ExtensionRegistryLite) registry); } - public interface FilterOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface FilterOrBuilder extends + // @@protoc_insertion_point(interface_extends:hbase.pb.Filter) + com.google.protobuf.MessageOrBuilder { - // required string name = 1; /** * <code>required string name = 1;</code> */ @@ -26,7 +32,6 @@ public final class FilterProtos { com.google.protobuf.ByteString getNameBytes(); - // optional bytes serialized_filter = 2; /** * <code>optional bytes serialized_filter = 2;</code> */ @@ -39,36 +44,29 @@ public final class FilterProtos { /** * Protobuf type {@code hbase.pb.Filter} */ - public static final class Filter extends - com.google.protobuf.GeneratedMessage - implements FilterOrBuilder { + public static final class Filter extends + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:hbase.pb.Filter) + FilterOrBuilder { // Use Filter.newBuilder() to construct. - private Filter(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + private Filter(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); - } - private Filter(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final Filter defaultInstance; - public static Filter getDefaultInstance() { - return defaultInstance; } - - public Filter getDefaultInstanceForType() { - return defaultInstance; + private Filter() { + name_ = ""; + serializedFilter_ = com.google.protobuf.ByteString.EMPTY; } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private Filter( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -88,8 +86,9 @@ public final class FilterProtos { break; } case 10: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; - name_ = input.readBytes(); + name_ = bs; break; } case 18: { @@ -103,7 +102,7 @@ public final class FilterProtos { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -114,32 +113,16 @@ public final class FilterProtos { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_Filter_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_Filter_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter.class, org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter.Builder.class); } - public static com.google.protobuf.Parser<Filter> PARSER = - new com.google.protobuf.AbstractParser<Filter>() { - public Filter parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new Filter(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser<Filter> getParserForType() { - return PARSER; - } - private int bitField0_; - // required string name = 1; public static final int NAME_FIELD_NUMBER = 1; - private java.lang.Object name_; + private volatile java.lang.Object name_; /** * <code>required string name = 1;</code> */ @@ -180,7 +163,6 @@ public final class FilterProtos { } } - // optional bytes serialized_filter = 2; public static final int SERIALIZED_FILTER_FIELD_NUMBER = 2; private com.google.protobuf.ByteString serializedFilter_; /** @@ -196,14 +178,11 @@ public final class FilterProtos { return serializedFilter_; } - private void initFields() { - name_ = ""; - serializedFilter_ = com.google.protobuf.ByteString.EMPTY; - } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; if (!hasName()) { memoizedIsInitialized = 0; @@ -215,43 +194,34 @@ public final class FilterProtos { public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getNameBytes()); + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, name_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeBytes(2, serializedFilter_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getNameBytes()); + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, name_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(2, serializedFilter_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; @@ -272,12 +242,10 @@ public final class FilterProtos { result = result && getSerializedFilter() .equals(other.getSerializedFilter()); } - result = result && - getUnknownFields().equals(other.getUnknownFields()); + result = result && unknownFields.equals(other.unknownFields); return result; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { @@ -293,7 +261,7 @@ public final class FilterProtos { hash = (37 * hash) + SERIALIZED_FILTER_FIELD_NUMBER; hash = (53 * hash) + getSerializedFilter().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } @@ -321,46 +289,57 @@ public final class FilterProtos { } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -368,14 +347,15 @@ public final class FilterProtos { * Protobuf type {@code hbase.pb.Filter} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder<Builder> - implements org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.FilterOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements + // @@protoc_insertion_point(builder_implements:hbase.pb.Filter) + org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.FilterOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_Filter_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_Filter_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -388,18 +368,15 @@ public final class FilterProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - public Builder clear() { super.clear(); name_ = ""; @@ -409,10 +386,6 @@ public final class FilterProtos { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_Filter_descriptor; @@ -447,6 +420,32 @@ public final class FilterProtos { return result; } + public Builder clone() { + return (Builder) super.clone(); + } + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + Object value) { + return (Builder) super.setField(field, value); + } + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return (Builder) super.clearField(field); + } + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return (Builder) super.clearOneof(oneof); + } + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, Object value) { + return (Builder) super.setRepeatedField(field, index, value); + } + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + Object value) { + return (Builder) super.addRepeatedField(field, value); + } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter)other); @@ -466,13 +465,13 @@ public final class FilterProtos { if (other.hasSerializedFilter()) { setSerializedFilter(other.getSerializedFilter()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } public final boolean isInitialized() { if (!hasName()) { - return false; } return true; @@ -487,7 +486,7 @@ public final class FilterProtos { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -497,7 +496,6 @@ public final class FilterProtos { } private int bitField0_; - // required string name = 1; private java.lang.Object name_ = ""; /** * <code>required string name = 1;</code> @@ -511,9 +509,12 @@ public final class FilterProtos { public java.lang.String getName() { java.lang.Object ref = name_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - name_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + name_ = s; + } return s; } else { return (java.lang.String) ref; @@ -571,7 +572,6 @@ public final class FilterProtos { return this; } - // optional bytes serialized_filter = 2; private com.google.protobuf.ByteString serializedFilter_ = com.google.protobuf.ByteString.EMPTY; /** * <code>optional bytes serialized_filter = 2;</code> @@ -606,22 +606,59 @@ public final class FilterProtos { onChanged(); return this; } + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:hbase.pb.Filter) } + // @@protoc_insertion_point(class_scope:hbase.pb.Filter) + private static final org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter DEFAULT_INSTANCE; static { - defaultInstance = new Filter(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter(); + } + + public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser<Filter> + PARSER = new com.google.protobuf.AbstractParser<Filter>() { + public Filter parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new Filter(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser<Filter> parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser<Filter> getParserForType() { + return PARSER; + } + + public org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.Filter getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:hbase.pb.Filter) } - public interface ColumnCountGetFilterOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface ColumnCountGetFilterOrBuilder extends + // @@protoc_insertion_point(interface_extends:hbase.pb.ColumnCountGetFilter) + com.google.protobuf.MessageOrBuilder { - // required int32 limit = 1; /** * <code>required int32 limit = 1;</code> */ @@ -634,36 +671,28 @@ public final class FilterProtos { /** * Protobuf type {@code hbase.pb.ColumnCountGetFilter} */ - public static final class ColumnCountGetFilter extends - com.google.protobuf.GeneratedMessage - implements ColumnCountGetFilterOrBuilder { + public static final class ColumnCountGetFilter extends + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:hbase.pb.ColumnCountGetFilter) + ColumnCountGetFilterOrBuilder { // Use ColumnCountGetFilter.newBuilder() to construct. - private ColumnCountGetFilter(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + private ColumnCountGetFilter(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); - } - private ColumnCountGetFilter(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final ColumnCountGetFilter defaultInstance; - public static ColumnCountGetFilter getDefaultInstance() { - return defaultInstance; } - - public ColumnCountGetFilter getDefaultInstanceForType() { - return defaultInstance; + private ColumnCountGetFilter() { + limit_ = 0; } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private ColumnCountGetFilter( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -693,7 +722,7 @@ public final class FilterProtos { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -704,30 +733,14 @@ public final class FilterProtos { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnCountGetFilter_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnCountGetFilter_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter.class, org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter.Builder.class); } - public static com.google.protobuf.Parser<ColumnCountGetFilter> PARSER = - new com.google.protobuf.AbstractParser<ColumnCountGetFilter>() { - public ColumnCountGetFilter parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new ColumnCountGetFilter(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser<ColumnCountGetFilter> getParserForType() { - return PARSER; - } - private int bitField0_; - // required int32 limit = 1; public static final int LIMIT_FIELD_NUMBER = 1; private int limit_; /** @@ -743,13 +756,11 @@ public final class FilterProtos { return limit_; } - private void initFields() { - limit_ = 0; - } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; if (!hasLimit()) { memoizedIsInitialized = 0; @@ -761,16 +772,14 @@ public final class FilterProtos { public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeInt32(1, limit_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; @@ -778,19 +787,13 @@ public final class FilterProtos { size += com.google.protobuf.CodedOutputStream .computeInt32Size(1, limit_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; @@ -806,12 +809,10 @@ public final class FilterProtos { result = result && (getLimit() == other.getLimit()); } - result = result && - getUnknownFields().equals(other.getUnknownFields()); + result = result && unknownFields.equals(other.unknownFields); return result; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { @@ -823,7 +824,7 @@ public final class FilterProtos { hash = (37 * hash) + LIMIT_FIELD_NUMBER; hash = (53 * hash) + getLimit(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } @@ -851,46 +852,57 @@ public final class FilterProtos { } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -898,14 +910,15 @@ public final class FilterProtos { * Protobuf type {@code hbase.pb.ColumnCountGetFilter} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder<Builder> - implements org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilterOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements + // @@protoc_insertion_point(builder_implements:hbase.pb.ColumnCountGetFilter) + org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilterOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnCountGetFilter_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnCountGetFilter_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -918,18 +931,15 @@ public final class FilterProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - public Builder clear() { super.clear(); limit_ = 0; @@ -937,10 +947,6 @@ public final class FilterProtos { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnCountGetFilter_descriptor; @@ -971,6 +977,32 @@ public final class FilterProtos { return result; } + public Builder clone() { + return (Builder) super.clone(); + } + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + Object value) { + return (Builder) super.setField(field, value); + } + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return (Builder) super.clearField(field); + } + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return (Builder) super.clearOneof(oneof); + } + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, Object value) { + return (Builder) super.setRepeatedField(field, index, value); + } + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + Object value) { + return (Builder) super.addRepeatedField(field, value); + } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter)other); @@ -985,13 +1017,13 @@ public final class FilterProtos { if (other.hasLimit()) { setLimit(other.getLimit()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } public final boolean isInitialized() { if (!hasLimit()) { - return false; } return true; @@ -1006,7 +1038,7 @@ public final class FilterProtos { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -1016,7 +1048,6 @@ public final class FilterProtos { } private int bitField0_; - // required int32 limit = 1; private int limit_ ; /** * <code>required int32 limit = 1;</code> @@ -1048,22 +1079,59 @@ public final class FilterProtos { onChanged(); return this; } + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:hbase.pb.ColumnCountGetFilter) } + // @@protoc_insertion_point(class_scope:hbase.pb.ColumnCountGetFilter) + private static final org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter DEFAULT_INSTANCE; static { - defaultInstance = new ColumnCountGetFilter(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter(); + } + + public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser<ColumnCountGetFilter> + PARSER = new com.google.protobuf.AbstractParser<ColumnCountGetFilter>() { + public ColumnCountGetFilter parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new ColumnCountGetFilter(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser<ColumnCountGetFilter> parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser<ColumnCountGetFilter> getParserForType() { + return PARSER; + } + + public org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnCountGetFilter getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:hbase.pb.ColumnCountGetFilter) } - public interface ColumnPaginationFilterOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface ColumnPaginationFilterOrBuilder extends + // @@protoc_insertion_point(interface_extends:hbase.pb.ColumnPaginationFilter) + com.google.protobuf.MessageOrBuilder { - // required int32 limit = 1; /** * <code>required int32 limit = 1;</code> */ @@ -1073,7 +1141,6 @@ public final class FilterProtos { */ int getLimit(); - // optional int32 offset = 2; /** * <code>optional int32 offset = 2;</code> */ @@ -1083,7 +1150,6 @@ public final class FilterProtos { */ int getOffset(); - // optional bytes column_offset = 3; /** * <code>optional bytes column_offset = 3;</code> */ @@ -1096,36 +1162,30 @@ public final class FilterProtos { /** * Protobuf type {@code hbase.pb.ColumnPaginationFilter} */ - public static final class ColumnPaginationFilter extends - com.google.protobuf.GeneratedMessage - implements ColumnPaginationFilterOrBuilder { + public static final class ColumnPaginationFilter extends + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:hbase.pb.ColumnPaginationFilter) + ColumnPaginationFilterOrBuilder { // Use ColumnPaginationFilter.newBuilder() to construct. - private ColumnPaginationFilter(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + private ColumnPaginationFilter(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private ColumnPaginationFilter(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final ColumnPaginationFilter defaultInstance; - public static ColumnPaginationFilter getDefaultInstance() { - return defaultInstance; - } - - public ColumnPaginationFilter getDefaultInstanceForType() { - return defaultInstance; + private ColumnPaginationFilter() { + limit_ = 0; + offset_ = 0; + columnOffset_ = com.google.protobuf.ByteString.EMPTY; } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private ColumnPaginationFilter( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -1165,7 +1225,7 @@ public final class FilterProtos { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -1176,30 +1236,14 @@ public final class FilterProtos { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPaginationFilter_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPaginationFilter_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter.class, org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter.Builder.class); } - public static com.google.protobuf.Parser<ColumnPaginationFilter> PARSER = - new com.google.protobuf.AbstractParser<ColumnPaginationFilter>() { - public ColumnPaginationFilter parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new ColumnPaginationFilter(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser<ColumnPaginationFilter> getParserForType() { - return PARSER; - } - private int bitField0_; - // required int32 limit = 1; public static final int LIMIT_FIELD_NUMBER = 1; private int limit_; /** @@ -1215,7 +1259,6 @@ public final class FilterProtos { return limit_; } - // optional int32 offset = 2; public static final int OFFSET_FIELD_NUMBER = 2; private int offset_; /** @@ -1231,7 +1274,6 @@ public final class FilterProtos { return offset_; } - // optional bytes column_offset = 3; public static final int COLUMN_OFFSET_FIELD_NUMBER = 3; private com.google.protobuf.ByteString columnOffset_; /** @@ -1247,15 +1289,11 @@ public final class FilterProtos { return columnOffset_; } - private void initFields() { - limit_ = 0; - offset_ = 0; - columnOffset_ = com.google.protobuf.ByteString.EMPTY; - } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; if (!hasLimit()) { memoizedIsInitialized = 0; @@ -1267,7 +1305,6 @@ public final class FilterProtos { public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeInt32(1, limit_); } @@ -1277,12 +1314,11 @@ public final class FilterProtos { if (((bitField0_ & 0x00000004) == 0x00000004)) { output.writeBytes(3, columnOffset_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; @@ -1298,19 +1334,13 @@ public final class FilterProtos { size += com.google.protobuf.CodedOutputStream .computeBytesSize(3, columnOffset_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; @@ -1336,12 +1366,10 @@ public final class FilterProtos { result = result && getColumnOffset() .equals(other.getColumnOffset()); } - result = result && - getUnknownFields().equals(other.getUnknownFields()); + result = result && unknownFields.equals(other.unknownFields); return result; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { @@ -1361,7 +1389,7 @@ public final class FilterProtos { hash = (37 * hash) + COLUMN_OFFSET_FIELD_NUMBER; hash = (53 * hash) + getColumnOffset().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } @@ -1389,46 +1417,57 @@ public final class FilterProtos { } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -1436,14 +1475,15 @@ public final class FilterProtos { * Protobuf type {@code hbase.pb.ColumnPaginationFilter} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder<Builder> - implements org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilterOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements + // @@protoc_insertion_point(builder_implements:hbase.pb.ColumnPaginationFilter) + org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilterOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPaginationFilter_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPaginationFilter_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -1456,18 +1496,15 @@ public final class FilterProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - public Builder clear() { super.clear(); limit_ = 0; @@ -1479,10 +1516,6 @@ public final class FilterProtos { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPaginationFilter_descriptor; @@ -1521,6 +1554,32 @@ public final class FilterProtos { return result; } + public Builder clone() { + return (Builder) super.clone(); + } + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + Object value) { + return (Builder) super.setField(field, value); + } + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return (Builder) super.clearField(field); + } + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return (Builder) super.clearOneof(oneof); + } + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, Object value) { + return (Builder) super.setRepeatedField(field, index, value); + } + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + Object value) { + return (Builder) super.addRepeatedField(field, value); + } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter)other); @@ -1541,13 +1600,13 @@ public final class FilterProtos { if (other.hasColumnOffset()) { setColumnOffset(other.getColumnOffset()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } public final boolean isInitialized() { if (!hasLimit()) { - return false; } return true; @@ -1562,7 +1621,7 @@ public final class FilterProtos { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -1572,7 +1631,6 @@ public final class FilterProtos { } private int bitField0_; - // required int32 limit = 1; private int limit_ ; /** * <code>required int32 limit = 1;</code> @@ -1605,7 +1663,6 @@ public final class FilterProtos { return this; } - // optional int32 offset = 2; private int offset_ ; /** * <code>optional int32 offset = 2;</code> @@ -1638,7 +1695,6 @@ public final class FilterProtos { return this; } - // optional bytes column_offset = 3; private com.google.protobuf.ByteString columnOffset_ = com.google.protobuf.ByteString.EMPTY; /** * <code>optional bytes column_offset = 3;</code> @@ -1673,22 +1729,59 @@ public final class FilterProtos { onChanged(); return this; } + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:hbase.pb.ColumnPaginationFilter) } + // @@protoc_insertion_point(class_scope:hbase.pb.ColumnPaginationFilter) + private static final org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter DEFAULT_INSTANCE; static { - defaultInstance = new ColumnPaginationFilter(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter(); + } + + public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser<ColumnPaginationFilter> + PARSER = new com.google.protobuf.AbstractParser<ColumnPaginationFilter>() { + public ColumnPaginationFilter parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new ColumnPaginationFilter(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser<ColumnPaginationFilter> parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser<ColumnPaginationFilter> getParserForType() { + return PARSER; + } + + public org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPaginationFilter getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:hbase.pb.ColumnPaginationFilter) } - public interface ColumnPrefixFilterOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface ColumnPrefixFilterOrBuilder extends + // @@protoc_insertion_point(interface_extends:hbase.pb.ColumnPrefixFilter) + com.google.protobuf.MessageOrBuilder { - // required bytes prefix = 1; /** * <code>required bytes prefix = 1;</code> */ @@ -1701,36 +1794,28 @@ public final class FilterProtos { /** * Protobuf type {@code hbase.pb.ColumnPrefixFilter} */ - public static final class ColumnPrefixFilter extends - com.google.protobuf.GeneratedMessage - implements ColumnPrefixFilterOrBuilder { + public static final class ColumnPrefixFilter extends + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:hbase.pb.ColumnPrefixFilter) + ColumnPrefixFilterOrBuilder { // Use ColumnPrefixFilter.newBuilder() to construct. - private ColumnPrefixFilter(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + private ColumnPrefixFilter(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); - } - private ColumnPrefixFilter(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final ColumnPrefixFilter defaultInstance; - public static ColumnPrefixFilter getDefaultInstance() { - return defaultInstance; } - - public ColumnPrefixFilter getDefaultInstanceForType() { - return defaultInstance; + private ColumnPrefixFilter() { + prefix_ = com.google.protobuf.ByteString.EMPTY; } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private ColumnPrefixFilter( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -1760,7 +1845,7 @@ public final class FilterProtos { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -1771,30 +1856,14 @@ public final class FilterProtos { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPrefixFilter_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPrefixFilter_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter.class, org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter.Builder.class); } - public static com.google.protobuf.Parser<ColumnPrefixFilter> PARSER = - new com.google.protobuf.AbstractParser<ColumnPrefixFilter>() { - public ColumnPrefixFilter parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new ColumnPrefixFilter(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser<ColumnPrefixFilter> getParserForType() { - return PARSER; - } - private int bitField0_; - // required bytes prefix = 1; public static final int PREFIX_FIELD_NUMBER = 1; private com.google.protobuf.ByteString prefix_; /** @@ -1810,13 +1879,11 @@ public final class FilterProtos { return prefix_; } - private void initFields() { - prefix_ = com.google.protobuf.ByteString.EMPTY; - } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; if (!hasPrefix()) { memoizedIsInitialized = 0; @@ -1828,16 +1895,14 @@ public final class FilterProtos { public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeBytes(1, prefix_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; @@ -1845,19 +1910,13 @@ public final class FilterProtos { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, prefix_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; @@ -1873,12 +1932,10 @@ public final class FilterProtos { result = result && getPrefix() .equals(other.getPrefix()); } - result = result && - getUnknownFields().equals(other.getUnknownFields()); + result = result && unknownFields.equals(other.unknownFields); return result; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { @@ -1890,7 +1947,7 @@ public final class FilterProtos { hash = (37 * hash) + PREFIX_FIELD_NUMBER; hash = (53 * hash) + getPrefix().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } @@ -1918,46 +1975,57 @@ public final class FilterProtos { } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -1965,14 +2033,15 @@ public final class FilterProtos { * Protobuf type {@code hbase.pb.ColumnPrefixFilter} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder<Builder> - implements org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilterOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements + // @@protoc_insertion_point(builder_implements:hbase.pb.ColumnPrefixFilter) + org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilterOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPrefixFilter_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPrefixFilter_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -1985,18 +2054,15 @@ public final class FilterProtos { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - public Builder clear() { super.clear(); prefix_ = com.google.protobuf.ByteString.EMPTY; @@ -2004,10 +2070,6 @@ public final class FilterProtos { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnPrefixFilter_descriptor; @@ -2038,6 +2100,32 @@ public final class FilterProtos { return result; } + public Builder clone() { + return (Builder) super.clone(); + } + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + Object value) { + return (Builder) super.setField(field, value); + } + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return (Builder) super.clearField(field); + } + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return (Builder) super.clearOneof(oneof); + } + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, Object value) { + return (Builder) super.setRepeatedField(field, index, value); + } + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + Object value) { + return (Builder) super.addRepeatedField(field, value); + } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter)other); @@ -2052,13 +2140,13 @@ public final class FilterProtos { if (other.hasPrefix()) { setPrefix(other.getPrefix()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } public final boolean isInitialized() { if (!hasPrefix()) { - return false; } return true; @@ -2073,7 +2161,7 @@ public final class FilterProtos { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -2083,7 +2171,6 @@ public final class FilterProtos { } private int bitField0_; - // required bytes prefix = 1; private com.google.protobuf.ByteString prefix_ = com.google.protobuf.ByteString.EMPTY; /** * <code>required bytes prefix = 1;</code> @@ -2118,32 +2205,68 @@ public final class FilterProtos { onChanged(); return this; } + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:hbase.pb.ColumnPrefixFilter) } + // @@protoc_insertion_point(class_scope:hbase.pb.ColumnPrefixFilter) + private static final org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter DEFAULT_INSTANCE; static { - defaultInstance = new ColumnPrefixFilter(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter(); } - // @@protoc_insertion_point(class_scope:hbase.pb.ColumnPrefixFilter) - } + public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter getDefaultInstance() { + return DEFAULT_INSTANCE; + } - public interface ColumnRangeFilterOrBuilder - extends com.google.protobuf.MessageOrBuilder { + @java.lang.Deprecated public static final com.google.protobuf.Parser<ColumnPrefixFilter> + PARSER = new com.google.protobuf.AbstractParser<ColumnPrefixFilter>() { + public ColumnPrefixFilter parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new ColumnPrefixFilter(input, extensionRegistry); + } + }; - // optional bytes min_column = 1; - /** - * <code>optional bytes min_column = 1;</code> - */ - boolean hasMinColumn(); - /** - * <code>optional bytes min_column = 1;</code> - */ + public static com.google.protobuf.Parser<ColumnPrefixFilter> parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser<ColumnPrefixFilter> getParserForType() { + return PARSER; + } + + public org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnPrefixFilter getDefaultInstanceForType() { + return DEFAULT_INSTANCE; + } + + } + + public interface ColumnRangeFilterOrBuilder extends + // @@protoc_insertion_point(interface_extends:hbase.pb.ColumnRangeFilter) + com.google.protobuf.MessageOrBuilder { + + /** + * <code>optional bytes min_column = 1;</code> + */ + boolean hasMinColumn(); + /** + * <code>optional bytes min_column = 1;</code> + */ com.google.protobuf.ByteString getMinColumn(); - // optional bool min_column_inclusive = 2; /** * <code>optional bool min_column_inclusive = 2;</code> */ @@ -2153,7 +2276,6 @@ public final class FilterProtos { */ boolean getMinColumnInclusive(); - // optional bytes max_column = 3; /** * <code>optional bytes max_column = 3;</code> */ @@ -2163,7 +2285,6 @@ public final class FilterProtos { */ com.google.protobuf.ByteString getMaxColumn(); - // optional bool max_column_inclusive = 4; /** * <code>optional bool max_column_inclusive = 4;</code> */ @@ -2176,36 +2297,31 @@ public final class FilterProtos { /** * Protobuf type {@code hbase.pb.ColumnRangeFilter} */ - public static final class ColumnRangeFilter extends - com.google.protobuf.GeneratedMessage - implements ColumnRangeFilterOrBuilder { + public static final class ColumnRangeFilter extends + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:hbase.pb.ColumnRangeFilter) + ColumnRangeFilterOrBuilder { // Use ColumnRangeFilter.newBuilder() to construct. - private ColumnRangeFilter(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + private ColumnRangeFilter(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private ColumnRangeFilter(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final ColumnRangeFilter defaultInstance; - public static ColumnRangeFilter getDefaultInstance() { - return defaultInstance; - } - - public ColumnRangeFilter getDefaultInstanceForType() { - return defaultInstance; + private ColumnRangeFilter() { + minColumn_ = com.google.protobuf.ByteString.EMPTY; + minColumnInclusive_ = false; + maxColumn_ = com.google.protobuf.ByteString.EMPTY; + maxColumnInclusive_ = false; } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private ColumnRangeFilter( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -2250,7 +2366,7 @@ public final class FilterProtos { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -2261,30 +2377,14 @@ public final class FilterProtos { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnRangeFilter_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.internal_static_hbase_pb_ColumnRangeFilter_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilter.class, org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilter.Builder.class); } - public static com.google.protobuf.Parser<ColumnRangeFilter> PARSER = - new com.google.protobuf.AbstractParser<ColumnRangeFilter>() { - public ColumnRangeFilter parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new ColumnRangeFilter(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser<ColumnRangeFilter> getParserForType() { - return PARSER; - } - private int bitField0_; - // optional bytes min_column = 1; public static final int MIN_COLUMN_FIELD_NUMBER = 1; private com.google.protobuf.ByteString minColumn_; /** @@ -2300,7 +2400,6 @@ public final class FilterProtos { return minColumn_; } - // optional bool min_column_inclusive = 2; public static final int MIN_COLUMN_INCLUSIVE_FIELD_NUMBER = 2; private boolean minColumnInclusive_; /** @@ -2316,7 +2415,6 @@ public final class FilterProtos { return minColumnInclusive_; } - // optional bytes max_column = 3; public static final int MAX_COLUMN_FIELD_NUMBER = 3; private com.google.protobuf.ByteString maxColumn_; /** @@ -2332,7 +2430,6 @@ public final class FilterProtos { return maxColumn_; } - // optional bool max_column_inclusive = 4; public static final int MAX_COLUMN_INCLUSIVE_FIELD_NUMBER = 4; private boolean maxColumnInclusive_; /** @@ -2348,16 +2445,11 @@ public final class FilterProtos { return maxColumnInclusive_; } - private void initFields() { - minColumn_ = com.google.protobuf.ByteString.EMPTY; - minColumnInclusive_ = false; - maxColumn_ = com.google.protobuf.ByteString.EMPTY; - maxColumnInclusive_ = false; - } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; @@ -2365,7 +2457,6 @@ public final class FilterProtos { public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeBytes(1, minColumn_); } @@ -2378,12 +2469,11 @@ public final class FilterProtos { if (((bitField0_ & 0x00000008) == 0x00000008)) { output.writeBool(4, maxColumnInclusive_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; @@ -2403,19 +2493,13 @@ public final class FilterProtos { size += com.google.protobuf.CodedOutputStream .computeBoolSize(4, maxColumnInclusive_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; @@ -2446,12 +2530,10 @@ public final class FilterProtos { result = result && (getMaxColumnInclusive() == other.getMaxColumnInclusive()); } - result = result && - getUnknownFields().equals(other.getUnknownFields()); + result = result && unknownFields.equals(other.unknownFields); return result; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { @@ -2465,7 +2547,8 @@ public final class FilterProtos { } if (hasMinColumnInclusive()) { hash = (37 * hash) + MIN_COLUMN_INCLUSIVE_FIELD_NUMBER; - hash = (53 * hash) + hashBoolean(getMinColumnInclusive()); + hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean( + getMinColumnInclusive()); } if (hasMaxColumn()) { hash = (37 * hash) + MAX_COLUMN_FIELD_NUMBER; @@ -2473,9 +2556,10 @@ public final class FilterProtos { } if (hasMaxColumnInclusive()) { hash = (37 * hash) + MAX_COLUMN_INCLUSIVE_FIELD_NUMBER; - hash = (53 * hash) + hashBoolean(getMaxColumnInclusive()); + hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean( + getMaxColumnInclusive()); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } @@ -2503,46 +2587,57 @@ public final class FilterProtos { } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilter parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilter parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilter parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilter parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilter parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilter parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilter prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -2550,14 +2645,15 @@ public final class FilterProtos { * Protobuf type {@code hbase.pb.ColumnRangeFilter} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder<Builder> - implements org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilterOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements + // @@protoc_insertion_point(builder_implements:hbase.pb.ColumnRangeFilter) + org.apache.hadoop.hbase.shaded.protobuf.generated.FilterProtos.ColumnRangeFilterOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.FilterP
<TRUNCATED>