http://git-wip-us.apache.org/repos/asf/hbase/blob/b4a729ed/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FSProtos.java
----------------------------------------------------------------------
diff --git
a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FSProtos.java
b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FSProtos.java
index 74d3f86..5b456c8 100644
---
a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FSProtos.java
+++
b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FSProtos.java
@@ -6,12 +6,18 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
public final class FSProtos {
private FSProtos() {}
public static void registerAllExtensions(
+ com.google.protobuf.ExtensionRegistryLite registry) {
+ }
+
+ public static void registerAllExtensions(
com.google.protobuf.ExtensionRegistry registry) {
+ registerAllExtensions(
+ (com.google.protobuf.ExtensionRegistryLite) registry);
}
- public interface HBaseVersionFileContentOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface HBaseVersionFileContentOrBuilder extends
+ //
@@protoc_insertion_point(interface_extends:hbase.pb.HBaseVersionFileContent)
+ com.google.protobuf.MessageOrBuilder {
- // required string version = 1;
/**
* <code>required string version = 1;</code>
*/
@@ -27,43 +33,35 @@ public final class FSProtos {
getVersionBytes();
}
/**
- * Protobuf type {@code hbase.pb.HBaseVersionFileContent}
- *
* <pre>
**
* The ${HBASE_ROOTDIR}/hbase.version file content
* </pre>
+ *
+ * Protobuf type {@code hbase.pb.HBaseVersionFileContent}
*/
- public static final class HBaseVersionFileContent extends
- com.google.protobuf.GeneratedMessage
- implements HBaseVersionFileContentOrBuilder {
+ public static final class HBaseVersionFileContent extends
+ com.google.protobuf.GeneratedMessageV3 implements
+ //
@@protoc_insertion_point(message_implements:hbase.pb.HBaseVersionFileContent)
+ HBaseVersionFileContentOrBuilder {
// Use HBaseVersionFileContent.newBuilder() to construct.
- private
HBaseVersionFileContent(com.google.protobuf.GeneratedMessage.Builder<?>
builder) {
+ private
HBaseVersionFileContent(com.google.protobuf.GeneratedMessageV3.Builder<?>
builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
}
- private HBaseVersionFileContent(boolean noInit) { this.unknownFields =
com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final HBaseVersionFileContent defaultInstance;
- public static HBaseVersionFileContent getDefaultInstance() {
- return defaultInstance;
- }
-
- public HBaseVersionFileContent getDefaultInstanceForType() {
- return defaultInstance;
+ private HBaseVersionFileContent() {
+ version_ = "";
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
+ getUnknownFields() {
return this.unknownFields;
}
private HBaseVersionFileContent(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
+ this();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
@@ -83,8 +81,9 @@ public final class FSProtos {
break;
}
case 10: {
+ com.google.protobuf.ByteString bs = input.readBytes();
bitField0_ |= 0x00000001;
- version_ = input.readBytes();
+ version_ = bs;
break;
}
}
@@ -93,7 +92,7 @@ public final class FSProtos {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
+ e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
@@ -104,32 +103,16 @@ public final class FSProtos {
return
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_HBaseVersionFileContent_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_HBaseVersionFileContent_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent.class,
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent.Builder.class);
}
- public static com.google.protobuf.Parser<HBaseVersionFileContent> PARSER =
- new com.google.protobuf.AbstractParser<HBaseVersionFileContent>() {
- public HBaseVersionFileContent parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new HBaseVersionFileContent(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<HBaseVersionFileContent>
getParserForType() {
- return PARSER;
- }
-
private int bitField0_;
- // required string version = 1;
public static final int VERSION_FIELD_NUMBER = 1;
- private java.lang.Object version_;
+ private volatile java.lang.Object version_;
/**
* <code>required string version = 1;</code>
*/
@@ -170,13 +153,11 @@ public final class FSProtos {
}
}
- private void initFields() {
- version_ = "";
- }
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
if (!hasVersion()) {
memoizedIsInitialized = 0;
@@ -188,36 +169,27 @@ public final class FSProtos {
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- output.writeBytes(1, getVersionBytes());
+ com.google.protobuf.GeneratedMessageV3.writeString(output, 1,
version_);
}
- getUnknownFields().writeTo(output);
+ unknownFields.writeTo(output);
}
- private int memoizedSerializedSize = -1;
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
- size += com.google.protobuf.CodedOutputStream
- .computeBytesSize(1, getVersionBytes());
+ size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1,
version_);
}
- size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ size += unknownFields.getSerializedSize();
+ memoizedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
- @java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
@@ -233,12 +205,10 @@ public final class FSProtos {
result = result && getVersion()
.equals(other.getVersion());
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
+ result = result && unknownFields.equals(other.unknownFields);
return result;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
@@ -250,7 +220,7 @@ public final class FSProtos {
hash = (37 * hash) + VERSION_FIELD_NUMBER;
hash = (53 * hash) + getVersion().hashCode();
}
- hash = (29 * hash) + getUnknownFields().hashCode();
+ hash = (29 * hash) + unknownFields.hashCode();
memoizedHashCode = hash;
return hash;
}
@@ -278,66 +248,78 @@ public final class FSProtos {
}
public static
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent
parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent
parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
public static
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent
parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
public static
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent
parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent
parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent
parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder
newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent
prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
- * Protobuf type {@code hbase.pb.HBaseVersionFileContent}
- *
* <pre>
**
* The ${HBASE_ROOTDIR}/hbase.version file content
* </pre>
+ *
+ * Protobuf type {@code hbase.pb.HBaseVersionFileContent}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContentOrBuilder
{
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ //
@@protoc_insertion_point(builder_implements:hbase.pb.HBaseVersionFileContent)
+
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContentOrBuilder
{
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_HBaseVersionFileContent_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_HBaseVersionFileContent_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -350,18 +332,15 @@ public final class FSProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ if (com.google.protobuf.GeneratedMessageV3
+ .alwaysUseFieldBuilders) {
}
}
- private static Builder create() {
- return new Builder();
- }
-
public Builder clear() {
super.clear();
version_ = "";
@@ -369,10 +348,6 @@ public final class FSProtos {
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_HBaseVersionFileContent_descriptor;
@@ -403,6 +378,32 @@ public final class FSProtos {
return result;
}
+ public Builder clone() {
+ return (Builder) super.clone();
+ }
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.setField(field, value);
+ }
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return (Builder) super.clearField(field);
+ }
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return (Builder) super.clearOneof(oneof);
+ }
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, Object value) {
+ return (Builder) super.setRepeatedField(field, index, value);
+ }
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.addRepeatedField(field, value);
+ }
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent)
{
return
mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent)other);
@@ -419,13 +420,13 @@ public final class FSProtos {
version_ = other.version_;
onChanged();
}
- this.mergeUnknownFields(other.getUnknownFields());
+ this.mergeUnknownFields(other.unknownFields);
+ onChanged();
return this;
}
public final boolean isInitialized() {
if (!hasVersion()) {
-
return false;
}
return true;
@@ -440,7 +441,7 @@ public final class FSProtos {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage =
(org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent)
e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
@@ -450,7 +451,6 @@ public final class FSProtos {
}
private int bitField0_;
- // required string version = 1;
private java.lang.Object version_ = "";
/**
* <code>required string version = 1;</code>
@@ -464,9 +464,12 @@ public final class FSProtos {
public java.lang.String getVersion() {
java.lang.Object ref = version_;
if (!(ref instanceof java.lang.String)) {
- java.lang.String s = ((com.google.protobuf.ByteString) ref)
- .toStringUtf8();
- version_ = s;
+ com.google.protobuf.ByteString bs =
+ (com.google.protobuf.ByteString) ref;
+ java.lang.String s = bs.toStringUtf8();
+ if (bs.isValidUtf8()) {
+ version_ = s;
+ }
return s;
} else {
return (java.lang.String) ref;
@@ -523,22 +526,59 @@ public final class FSProtos {
onChanged();
return this;
}
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
//
@@protoc_insertion_point(builder_scope:hbase.pb.HBaseVersionFileContent)
}
+ // @@protoc_insertion_point(class_scope:hbase.pb.HBaseVersionFileContent)
+ private static final
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent
DEFAULT_INSTANCE;
static {
- defaultInstance = new HBaseVersionFileContent(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent();
+ }
+
+ public static
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent
getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
+
+ @java.lang.Deprecated public static final
com.google.protobuf.Parser<HBaseVersionFileContent>
+ PARSER = new
com.google.protobuf.AbstractParser<HBaseVersionFileContent>() {
+ public HBaseVersionFileContent parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new HBaseVersionFileContent(input, extensionRegistry);
+ }
+ };
+
+ public static com.google.protobuf.Parser<HBaseVersionFileContent> parser()
{
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<HBaseVersionFileContent>
getParserForType() {
+ return PARSER;
+ }
+
+ public
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent
getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
}
- // @@protoc_insertion_point(class_scope:hbase.pb.HBaseVersionFileContent)
}
- public interface ReferenceOrBuilder
- extends com.google.protobuf.MessageOrBuilder {
+ public interface ReferenceOrBuilder extends
+ // @@protoc_insertion_point(interface_extends:hbase.pb.Reference)
+ com.google.protobuf.MessageOrBuilder {
- // required bytes splitkey = 1;
/**
* <code>required bytes splitkey = 1;</code>
*/
@@ -548,7 +588,6 @@ public final class FSProtos {
*/
com.google.protobuf.ByteString getSplitkey();
- // required .hbase.pb.Reference.Range range = 2;
/**
* <code>required .hbase.pb.Reference.Range range = 2;</code>
*/
@@ -559,43 +598,36 @@ public final class FSProtos {
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range
getRange();
}
/**
- * Protobuf type {@code hbase.pb.Reference}
- *
* <pre>
**
* Reference file content used when we split an hfile under a region.
* </pre>
+ *
+ * Protobuf type {@code hbase.pb.Reference}
*/
- public static final class Reference extends
- com.google.protobuf.GeneratedMessage
- implements ReferenceOrBuilder {
+ public static final class Reference extends
+ com.google.protobuf.GeneratedMessageV3 implements
+ // @@protoc_insertion_point(message_implements:hbase.pb.Reference)
+ ReferenceOrBuilder {
// Use Reference.newBuilder() to construct.
- private Reference(com.google.protobuf.GeneratedMessage.Builder<?> builder)
{
+ private Reference(com.google.protobuf.GeneratedMessageV3.Builder<?>
builder) {
super(builder);
- this.unknownFields = builder.getUnknownFields();
- }
- private Reference(boolean noInit) { this.unknownFields =
com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
- private static final Reference defaultInstance;
- public static Reference getDefaultInstance() {
- return defaultInstance;
}
-
- public Reference getDefaultInstanceForType() {
- return defaultInstance;
+ private Reference() {
+ splitkey_ = com.google.protobuf.ByteString.EMPTY;
+ range_ = 0;
}
- private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
- getUnknownFields() {
+ getUnknownFields() {
return this.unknownFields;
}
private Reference(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
- initFields();
+ this();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
@@ -626,7 +658,7 @@ public final class FSProtos {
unknownFields.mergeVarintField(2, rawValue);
} else {
bitField0_ |= 0x00000002;
- range_ = value;
+ range_ = rawValue;
}
break;
}
@@ -636,7 +668,7 @@ public final class FSProtos {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
- e.getMessage()).setUnfinishedMessage(this);
+ e).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
@@ -647,28 +679,13 @@ public final class FSProtos {
return
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_Reference_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_Reference_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.class,
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Builder.class);
}
- public static com.google.protobuf.Parser<Reference> PARSER =
- new com.google.protobuf.AbstractParser<Reference>() {
- public Reference parsePartialFrom(
- com.google.protobuf.CodedInputStream input,
- com.google.protobuf.ExtensionRegistryLite extensionRegistry)
- throws com.google.protobuf.InvalidProtocolBufferException {
- return new Reference(input, extensionRegistry);
- }
- };
-
- @java.lang.Override
- public com.google.protobuf.Parser<Reference> getParserForType() {
- return PARSER;
- }
-
/**
* Protobuf enum {@code hbase.pb.Reference.Range}
*/
@@ -677,11 +694,11 @@ public final class FSProtos {
/**
* <code>TOP = 0;</code>
*/
- TOP(0, 0),
+ TOP(0),
/**
* <code>BOTTOM = 1;</code>
*/
- BOTTOM(1, 1),
+ BOTTOM(1),
;
/**
@@ -694,9 +711,19 @@ public final class FSProtos {
public static final int BOTTOM_VALUE = 1;
- public final int getNumber() { return value; }
+ public final int getNumber() {
+ return value;
+ }
+ /**
+ * @deprecated Use {@link #forNumber(int)} instead.
+ */
+ @java.lang.Deprecated
public static Range valueOf(int value) {
+ return forNumber(value);
+ }
+
+ public static Range forNumber(int value) {
switch (value) {
case 0: return TOP;
case 1: return BOTTOM;
@@ -708,17 +735,17 @@ public final class FSProtos {
internalGetValueMap() {
return internalValueMap;
}
- private static com.google.protobuf.Internal.EnumLiteMap<Range>
- internalValueMap =
+ private static final com.google.protobuf.Internal.EnumLiteMap<
+ Range> internalValueMap =
new com.google.protobuf.Internal.EnumLiteMap<Range>() {
public Range findValueByNumber(int number) {
- return Range.valueOf(number);
+ return Range.forNumber(number);
}
};
public final com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
- return getDescriptor().getValues().get(index);
+ return getDescriptor().getValues().get(ordinal());
}
public final com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
@@ -740,11 +767,9 @@ public final class FSProtos {
return VALUES[desc.getIndex()];
}
- private final int index;
private final int value;
- private Range(int index, int value) {
- this.index = index;
+ private Range(int value) {
this.value = value;
}
@@ -752,7 +777,6 @@ public final class FSProtos {
}
private int bitField0_;
- // required bytes splitkey = 1;
public static final int SPLITKEY_FIELD_NUMBER = 1;
private com.google.protobuf.ByteString splitkey_;
/**
@@ -768,9 +792,8 @@ public final class FSProtos {
return splitkey_;
}
- // required .hbase.pb.Reference.Range range = 2;
public static final int RANGE_FIELD_NUMBER = 2;
- private
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range
range_;
+ private int range_;
/**
* <code>required .hbase.pb.Reference.Range range = 2;</code>
*/
@@ -781,17 +804,15 @@ public final class FSProtos {
* <code>required .hbase.pb.Reference.Range range = 2;</code>
*/
public
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range
getRange() {
- return range_;
+
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range
result =
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range.valueOf(range_);
+ return result == null ?
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range.TOP
: result;
}
- private void initFields() {
- splitkey_ = com.google.protobuf.ByteString.EMPTY;
- range_ =
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range.TOP;
- }
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
- if (isInitialized != -1) return isInitialized == 1;
+ if (isInitialized == 1) return true;
+ if (isInitialized == 0) return false;
if (!hasSplitkey()) {
memoizedIsInitialized = 0;
@@ -807,19 +828,17 @@ public final class FSProtos {
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
- getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, splitkey_);
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
- output.writeEnum(2, range_.getNumber());
+ output.writeEnum(2, range_);
}
- getUnknownFields().writeTo(output);
+ unknownFields.writeTo(output);
}
- private int memoizedSerializedSize = -1;
public int getSerializedSize() {
- int size = memoizedSerializedSize;
+ int size = memoizedSize;
if (size != -1) return size;
size = 0;
@@ -829,21 +848,15 @@ public final class FSProtos {
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
size += com.google.protobuf.CodedOutputStream
- .computeEnumSize(2, range_.getNumber());
+ .computeEnumSize(2, range_);
}
- size += getUnknownFields().getSerializedSize();
- memoizedSerializedSize = size;
+ size += unknownFields.getSerializedSize();
+ memoizedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
- protected java.lang.Object writeReplace()
- throws java.io.ObjectStreamException {
- return super.writeReplace();
- }
-
- @java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
@@ -861,15 +874,12 @@ public final class FSProtos {
}
result = result && (hasRange() == other.hasRange());
if (hasRange()) {
- result = result &&
- (getRange() == other.getRange());
+ result = result && range_ == other.range_;
}
- result = result &&
- getUnknownFields().equals(other.getUnknownFields());
+ result = result && unknownFields.equals(other.unknownFields);
return result;
}
- private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
@@ -883,9 +893,9 @@ public final class FSProtos {
}
if (hasRange()) {
hash = (37 * hash) + RANGE_FIELD_NUMBER;
- hash = (53 * hash) + hashEnum(getRange());
+ hash = (53 * hash) + range_;
}
- hash = (29 * hash) + getUnknownFields().hashCode();
+ hash = (29 * hash) + unknownFields.hashCode();
memoizedHashCode = hash;
return hash;
}
@@ -913,66 +923,78 @@ public final class FSProtos {
}
public static
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference
parseFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
public static
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference
parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input);
}
public static
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference
parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseDelimitedFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
- return PARSER.parseFrom(input);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input);
}
public static
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
- return PARSER.parseFrom(input, extensionRegistry);
+ return com.google.protobuf.GeneratedMessageV3
+ .parseWithIOException(PARSER, input, extensionRegistry);
}
- public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
+ public static Builder newBuilder() {
+ return DEFAULT_INSTANCE.toBuilder();
+ }
public static Builder
newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference
prototype) {
- return newBuilder().mergeFrom(prototype);
+ return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+ }
+ public Builder toBuilder() {
+ return this == DEFAULT_INSTANCE
+ ? new Builder() : new Builder().mergeFrom(this);
}
- public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
- * Protobuf type {@code hbase.pb.Reference}
- *
* <pre>
**
* Reference file content used when we split an hfile under a region.
* </pre>
+ *
+ * Protobuf type {@code hbase.pb.Reference}
*/
public static final class Builder extends
- com.google.protobuf.GeneratedMessage.Builder<Builder>
- implements
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.ReferenceOrBuilder {
+ com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+ // @@protoc_insertion_point(builder_implements:hbase.pb.Reference)
+
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.ReferenceOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_Reference_descriptor;
}
- protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_Reference_fieldAccessorTable
.ensureFieldAccessorsInitialized(
@@ -985,31 +1007,24 @@ public final class FSProtos {
}
private Builder(
- com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+ com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
- if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+ if (com.google.protobuf.GeneratedMessageV3
+ .alwaysUseFieldBuilders) {
}
}
- private static Builder create() {
- return new Builder();
- }
-
public Builder clear() {
super.clear();
splitkey_ = com.google.protobuf.ByteString.EMPTY;
bitField0_ = (bitField0_ & ~0x00000001);
- range_ =
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range.TOP;
+ range_ = 0;
bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
- public Builder clone() {
- return create().mergeFrom(buildPartial());
- }
-
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_Reference_descriptor;
@@ -1044,6 +1059,32 @@ public final class FSProtos {
return result;
}
+ public Builder clone() {
+ return (Builder) super.clone();
+ }
+ public Builder setField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.setField(field, value);
+ }
+ public Builder clearField(
+ com.google.protobuf.Descriptors.FieldDescriptor field) {
+ return (Builder) super.clearField(field);
+ }
+ public Builder clearOneof(
+ com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+ return (Builder) super.clearOneof(oneof);
+ }
+ public Builder setRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ int index, Object value) {
+ return (Builder) super.setRepeatedField(field, index, value);
+ }
+ public Builder addRepeatedField(
+ com.google.protobuf.Descriptors.FieldDescriptor field,
+ Object value) {
+ return (Builder) super.addRepeatedField(field, value);
+ }
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference) {
return
mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference)other);
@@ -1061,17 +1102,16 @@ public final class FSProtos {
if (other.hasRange()) {
setRange(other.getRange());
}
- this.mergeUnknownFields(other.getUnknownFields());
+ this.mergeUnknownFields(other.unknownFields);
+ onChanged();
return this;
}
public final boolean isInitialized() {
if (!hasSplitkey()) {
-
return false;
}
if (!hasRange()) {
-
return false;
}
return true;
@@ -1086,7 +1126,7 @@ public final class FSProtos {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage =
(org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference)
e.getUnfinishedMessage();
- throw e;
+ throw e.unwrapIOException();
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
@@ -1096,7 +1136,6 @@ public final class FSProtos {
}
private int bitField0_;
- // required bytes splitkey = 1;
private com.google.protobuf.ByteString splitkey_ =
com.google.protobuf.ByteString.EMPTY;
/**
* <code>required bytes splitkey = 1;</code>
@@ -1132,8 +1171,7 @@ public final class FSProtos {
return this;
}
- // required .hbase.pb.Reference.Range range = 2;
- private
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range
range_ =
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range.TOP;
+ private int range_ = 0;
/**
* <code>required .hbase.pb.Reference.Range range = 2;</code>
*/
@@ -1144,7 +1182,8 @@ public final class FSProtos {
* <code>required .hbase.pb.Reference.Range range = 2;</code>
*/
public
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range
getRange() {
- return range_;
+
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range
result =
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range.valueOf(range_);
+ return result == null ?
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range.TOP
: result;
}
/**
* <code>required .hbase.pb.Reference.Range range = 2;</code>
@@ -1154,7 +1193,7 @@ public final class FSProtos {
throw new NullPointerException();
}
bitField0_ |= 0x00000002;
- range_ = value;
+ range_ = value.getNumber();
onChanged();
return this;
}
@@ -1163,38 +1202,75 @@ public final class FSProtos {
*/
public Builder clearRange() {
bitField0_ = (bitField0_ & ~0x00000002);
- range_ =
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range.TOP;
+ range_ = 0;
onChanged();
return this;
}
+ public final Builder setUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.setUnknownFields(unknownFields);
+ }
+
+ public final Builder mergeUnknownFields(
+ final com.google.protobuf.UnknownFieldSet unknownFields) {
+ return super.mergeUnknownFields(unknownFields);
+ }
+
// @@protoc_insertion_point(builder_scope:hbase.pb.Reference)
}
+ // @@protoc_insertion_point(class_scope:hbase.pb.Reference)
+ private static final
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference
DEFAULT_INSTANCE;
static {
- defaultInstance = new Reference(true);
- defaultInstance.initFields();
+ DEFAULT_INSTANCE = new
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference();
+ }
+
+ public static
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference
getDefaultInstance() {
+ return DEFAULT_INSTANCE;
+ }
+
+ @java.lang.Deprecated public static final
com.google.protobuf.Parser<Reference>
+ PARSER = new com.google.protobuf.AbstractParser<Reference>() {
+ public Reference parsePartialFrom(
+ com.google.protobuf.CodedInputStream input,
+ com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+ throws com.google.protobuf.InvalidProtocolBufferException {
+ return new Reference(input, extensionRegistry);
+ }
+ };
+
+ public static com.google.protobuf.Parser<Reference> parser() {
+ return PARSER;
+ }
+
+ @java.lang.Override
+ public com.google.protobuf.Parser<Reference> getParserForType() {
+ return PARSER;
+ }
+
+ public
org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference
getDefaultInstanceForType() {
+ return DEFAULT_INSTANCE;
}
- // @@protoc_insertion_point(class_scope:hbase.pb.Reference)
}
- private static com.google.protobuf.Descriptors.Descriptor
+ private static final com.google.protobuf.Descriptors.Descriptor
internal_static_hbase_pb_HBaseVersionFileContent_descriptor;
- private static
- com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ private static final
+ com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hbase_pb_HBaseVersionFileContent_fieldAccessorTable;
- private static com.google.protobuf.Descriptors.Descriptor
+ private static final com.google.protobuf.Descriptors.Descriptor
internal_static_hbase_pb_Reference_descriptor;
- private static
- com.google.protobuf.GeneratedMessage.FieldAccessorTable
+ private static final
+ com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hbase_pb_Reference_fieldAccessorTable;
public static com.google.protobuf.Descriptors.FileDescriptor
getDescriptor() {
return descriptor;
}
- private static com.google.protobuf.Descriptors.FileDescriptor
+ private static com.google.protobuf.Descriptors.FileDescriptor
descriptor;
static {
java.lang.String[] descriptorData = {
@@ -1206,29 +1282,29 @@ public final class FSProtos {
"d.protobuf.generatedB\010FSProtosH\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner
assigner =
- new
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
- public com.google.protobuf.ExtensionRegistry assignDescriptors(
- com.google.protobuf.Descriptors.FileDescriptor root) {
- descriptor = root;
- internal_static_hbase_pb_HBaseVersionFileContent_descriptor =
- getDescriptor().getMessageTypes().get(0);
- internal_static_hbase_pb_HBaseVersionFileContent_fieldAccessorTable
= new
- com.google.protobuf.GeneratedMessage.FieldAccessorTable(
- internal_static_hbase_pb_HBaseVersionFileContent_descriptor,
- new java.lang.String[] { "Version", });
- internal_static_hbase_pb_Reference_descriptor =
- getDescriptor().getMessageTypes().get(1);
- internal_static_hbase_pb_Reference_fieldAccessorTable = new
- com.google.protobuf.GeneratedMessage.FieldAccessorTable(
- internal_static_hbase_pb_Reference_descriptor,
- new java.lang.String[] { "Splitkey", "Range", });
- return null;
- }
- };
+ new com.google.protobuf.Descriptors.FileDescriptor.
InternalDescriptorAssigner() {
+ public com.google.protobuf.ExtensionRegistry assignDescriptors(
+ com.google.protobuf.Descriptors.FileDescriptor root) {
+ descriptor = root;
+ return null;
+ }
+ };
com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
new com.google.protobuf.Descriptors.FileDescriptor[] {
}, assigner);
+ internal_static_hbase_pb_HBaseVersionFileContent_descriptor =
+ getDescriptor().getMessageTypes().get(0);
+ internal_static_hbase_pb_HBaseVersionFileContent_fieldAccessorTable = new
+ com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
+ internal_static_hbase_pb_HBaseVersionFileContent_descriptor,
+ new java.lang.String[] { "Version", });
+ internal_static_hbase_pb_Reference_descriptor =
+ getDescriptor().getMessageTypes().get(1);
+ internal_static_hbase_pb_Reference_fieldAccessorTable = new
+ com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
+ internal_static_hbase_pb_Reference_descriptor,
+ new java.lang.String[] { "Splitkey", "Range", });
}
// @@protoc_insertion_point(outer_class_scope)