http://git-wip-us.apache.org/repos/asf/hbase/blob/bf4fcc30/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterProtos.java ---------------------------------------------------------------------- diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterProtos.java index 073eba9..043d549 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterProtos.java @@ -8,88 +8,6 @@ public final class MasterProtos { public static void registerAllExtensions( com.google.protobuf.ExtensionRegistry registry) { } - /** - * Protobuf enum {@code hbase.pb.MasterSwitchType} - */ - public enum MasterSwitchType - implements com.google.protobuf.ProtocolMessageEnum { - /** - * <code>SPLIT = 0;</code> - */ - SPLIT(0, 0), - /** - * <code>MERGE = 1;</code> - */ - MERGE(1, 1), - ; - - /** - * <code>SPLIT = 0;</code> - */ - public static final int SPLIT_VALUE = 0; - /** - * <code>MERGE = 1;</code> - */ - public static final int MERGE_VALUE = 1; - - - public final int getNumber() { return value; } - - public static MasterSwitchType valueOf(int value) { - switch (value) { - case 0: return SPLIT; - case 1: return MERGE; - default: return null; - } - } - - public static com.google.protobuf.Internal.EnumLiteMap<MasterSwitchType> - internalGetValueMap() { - return internalValueMap; - } - private static com.google.protobuf.Internal.EnumLiteMap<MasterSwitchType> - internalValueMap = - new com.google.protobuf.Internal.EnumLiteMap<MasterSwitchType>() { - public MasterSwitchType findValueByNumber(int number) { - return MasterSwitchType.valueOf(number); - } - }; - - public final com.google.protobuf.Descriptors.EnumValueDescriptor - getValueDescriptor() { - return getDescriptor().getValues().get(index); - } - public final com.google.protobuf.Descriptors.EnumDescriptor - getDescriptorForType() { - return getDescriptor(); - } - public static final com.google.protobuf.Descriptors.EnumDescriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.getDescriptor().getEnumTypes().get(0); - } - - private static final MasterSwitchType[] VALUES = values(); - - public static MasterSwitchType valueOf( - com.google.protobuf.Descriptors.EnumValueDescriptor desc) { - if (desc.getType() != getDescriptor()) { - throw new java.lang.IllegalArgumentException( - "EnumValueDescriptor is not for this type."); - } - return VALUES[desc.getIndex()]; - } - - private final int index; - private final int value; - - private MasterSwitchType(int index, int value) { - this.index = index; - this.value = value; - } - - // @@protoc_insertion_point(enum_scope:hbase.pb.MasterSwitchType) - } - public interface AddColumnRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { @@ -28846,62 +28764,28 @@ public final class MasterProtos { // @@protoc_insertion_point(class_scope:hbase.pb.IsBalancerEnabledResponse) } - public interface SetSplitOrMergeEnabledRequestOrBuilder + public interface NormalizeRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - - // required bool enabled = 1; - /** - * <code>required bool enabled = 1;</code> - */ - boolean hasEnabled(); - /** - * <code>required bool enabled = 1;</code> - */ - boolean getEnabled(); - - // optional bool synchronous = 2; - /** - * <code>optional bool synchronous = 2;</code> - */ - boolean hasSynchronous(); - /** - * <code>optional bool synchronous = 2;</code> - */ - boolean getSynchronous(); - - // repeated .hbase.pb.MasterSwitchType switch_types = 3; - /** - * <code>repeated .hbase.pb.MasterSwitchType switch_types = 3;</code> - */ - java.util.List<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType> getSwitchTypesList(); - /** - * <code>repeated .hbase.pb.MasterSwitchType switch_types = 3;</code> - */ - int getSwitchTypesCount(); - /** - * <code>repeated .hbase.pb.MasterSwitchType switch_types = 3;</code> - */ - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType getSwitchTypes(int index); } /** - * Protobuf type {@code hbase.pb.SetSplitOrMergeEnabledRequest} + * Protobuf type {@code hbase.pb.NormalizeRequest} */ - public static final class SetSplitOrMergeEnabledRequest extends + public static final class NormalizeRequest extends com.google.protobuf.GeneratedMessage - implements SetSplitOrMergeEnabledRequestOrBuilder { - // Use SetSplitOrMergeEnabledRequest.newBuilder() to construct. - private SetSplitOrMergeEnabledRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + implements NormalizeRequestOrBuilder { + // Use NormalizeRequest.newBuilder() to construct. + private NormalizeRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private SetSplitOrMergeEnabledRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private NormalizeRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - private static final SetSplitOrMergeEnabledRequest defaultInstance; - public static SetSplitOrMergeEnabledRequest getDefaultInstance() { + private static final NormalizeRequest defaultInstance; + public static NormalizeRequest getDefaultInstance() { return defaultInstance; } - public SetSplitOrMergeEnabledRequest getDefaultInstanceForType() { + public NormalizeRequest getDefaultInstanceForType() { return defaultInstance; } @@ -28911,12 +28795,11 @@ public final class MasterProtos { getUnknownFields() { return this.unknownFields; } - private SetSplitOrMergeEnabledRequest( + private NormalizeRequest( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); - int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { @@ -28934,49 +28817,6 @@ public final class MasterProtos { } break; } - case 8: { - bitField0_ |= 0x00000001; - enabled_ = input.readBool(); - break; - } - case 16: { - bitField0_ |= 0x00000002; - synchronous_ = input.readBool(); - break; - } - case 24: { - int rawValue = input.readEnum(); - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType value = org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType.valueOf(rawValue); - if (value == null) { - unknownFields.mergeVarintField(3, rawValue); - } else { - if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) { - switchTypes_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType>(); - mutable_bitField0_ |= 0x00000004; - } - switchTypes_.add(value); - } - break; - } - case 26: { - int length = input.readRawVarint32(); - int oldLimit = input.pushLimit(length); - while(input.getBytesUntilLimit() > 0) { - int rawValue = input.readEnum(); - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType value = org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType.valueOf(rawValue); - if (value == null) { - unknownFields.mergeVarintField(3, rawValue); - } else { - if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) { - switchTypes_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType>(); - mutable_bitField0_ |= 0x00000004; - } - switchTypes_.add(value); - } - } - input.popLimit(oldLimit); - break; - } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { @@ -28985,109 +28825,44 @@ public final class MasterProtos { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { - if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) { - switchTypes_ = java.util.Collections.unmodifiableList(switchTypes_); - } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledRequest_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeRequest_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledRequest_fieldAccessorTable + return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest.Builder.class); + org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest.Builder.class); } - public static com.google.protobuf.Parser<SetSplitOrMergeEnabledRequest> PARSER = - new com.google.protobuf.AbstractParser<SetSplitOrMergeEnabledRequest>() { - public SetSplitOrMergeEnabledRequest parsePartialFrom( + public static com.google.protobuf.Parser<NormalizeRequest> PARSER = + new com.google.protobuf.AbstractParser<NormalizeRequest>() { + public NormalizeRequest parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - return new SetSplitOrMergeEnabledRequest(input, extensionRegistry); + return new NormalizeRequest(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser<SetSplitOrMergeEnabledRequest> getParserForType() { + public com.google.protobuf.Parser<NormalizeRequest> getParserForType() { return PARSER; } - private int bitField0_; - // required bool enabled = 1; - public static final int ENABLED_FIELD_NUMBER = 1; - private boolean enabled_; - /** - * <code>required bool enabled = 1;</code> - */ - public boolean hasEnabled() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * <code>required bool enabled = 1;</code> - */ - public boolean getEnabled() { - return enabled_; - } - - // optional bool synchronous = 2; - public static final int SYNCHRONOUS_FIELD_NUMBER = 2; - private boolean synchronous_; - /** - * <code>optional bool synchronous = 2;</code> - */ - public boolean hasSynchronous() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * <code>optional bool synchronous = 2;</code> - */ - public boolean getSynchronous() { - return synchronous_; - } - - // repeated .hbase.pb.MasterSwitchType switch_types = 3; - public static final int SWITCH_TYPES_FIELD_NUMBER = 3; - private java.util.List<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType> switchTypes_; - /** - * <code>repeated .hbase.pb.MasterSwitchType switch_types = 3;</code> - */ - public java.util.List<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType> getSwitchTypesList() { - return switchTypes_; - } - /** - * <code>repeated .hbase.pb.MasterSwitchType switch_types = 3;</code> - */ - public int getSwitchTypesCount() { - return switchTypes_.size(); - } - /** - * <code>repeated .hbase.pb.MasterSwitchType switch_types = 3;</code> - */ - public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType getSwitchTypes(int index) { - return switchTypes_.get(index); - } - private void initFields() { - enabled_ = false; - synchronous_ = false; - switchTypes_ = java.util.Collections.emptyList(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; - if (!hasEnabled()) { - memoizedIsInitialized = 0; - return false; - } memoizedIsInitialized = 1; return true; } @@ -29095,15 +28870,6 @@ public final class MasterProtos { public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBool(1, enabled_); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBool(2, synchronous_); - } - for (int i = 0; i < switchTypes_.size(); i++) { - output.writeEnum(3, switchTypes_.get(i).getNumber()); - } getUnknownFields().writeTo(output); } @@ -29113,23 +28879,6 @@ public final class MasterProtos { if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBoolSize(1, enabled_); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeBoolSize(2, synchronous_); - } - { - int dataSize = 0; - for (int i = 0; i < switchTypes_.size(); i++) { - dataSize += com.google.protobuf.CodedOutputStream - .computeEnumSizeNoTag(switchTypes_.get(i).getNumber()); - } - size += dataSize; - size += 1 * switchTypes_.size(); - } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; @@ -29147,24 +28896,12 @@ public final class MasterProtos { if (obj == this) { return true; } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest)) { + if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest)) { return super.equals(obj); } - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest) obj; + org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest) obj; boolean result = true; - result = result && (hasEnabled() == other.hasEnabled()); - if (hasEnabled()) { - result = result && (getEnabled() - == other.getEnabled()); - } - result = result && (hasSynchronous() == other.hasSynchronous()); - if (hasSynchronous()) { - result = result && (getSynchronous() - == other.getSynchronous()); - } - result = result && getSwitchTypesList() - .equals(other.getSwitchTypesList()); result = result && getUnknownFields().equals(other.getUnknownFields()); return result; @@ -29178,70 +28915,58 @@ public final class MasterProtos { } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasEnabled()) { - hash = (37 * hash) + ENABLED_FIELD_NUMBER; - hash = (53 * hash) + hashBoolean(getEnabled()); - } - if (hasSynchronous()) { - hash = (37 * hash) + SYNCHRONOUS_FIELD_NUMBER; - hash = (53 * hash) + hashBoolean(getSynchronous()); - } - if (getSwitchTypesCount() > 0) { - hash = (37 * hash) + SWITCH_TYPES_FIELD_NUMBER; - hash = (53 * hash) + hashEnumList(getSwitchTypesList()); - } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseFrom(byte[] data) + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseDelimitedFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseDelimitedFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -29250,7 +28975,7 @@ public final class MasterProtos { public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest prototype) { + public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @@ -29262,24 +28987,24 @@ public final class MasterProtos { return builder; } /** - * Protobuf type {@code hbase.pb.SetSplitOrMergeEnabledRequest} + * Protobuf type {@code hbase.pb.NormalizeRequest} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> - implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequestOrBuilder { + implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequestOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledRequest_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeRequest_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledRequest_fieldAccessorTable + return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest.Builder.class); + org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest.Builder.class); } - // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest.newBuilder() + // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } @@ -29299,12 +29024,6 @@ public final class MasterProtos { public Builder clear() { super.clear(); - enabled_ = false; - bitField0_ = (bitField0_ & ~0x00000001); - synchronous_ = false; - bitField0_ = (bitField0_ & ~0x00000002); - switchTypes_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000004); return this; } @@ -29314,79 +29033,43 @@ public final class MasterProtos { public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledRequest_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeRequest_descriptor; } - public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest.getDefaultInstance(); + public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest getDefaultInstanceForType() { + return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest.getDefaultInstance(); } - public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest build() { - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest result = buildPartial(); + public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest build() { + org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } - public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest(this); - int from_bitField0_ = bitField0_; - int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { - to_bitField0_ |= 0x00000001; - } - result.enabled_ = enabled_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { - to_bitField0_ |= 0x00000002; - } - result.synchronous_ = synchronous_; - if (((bitField0_ & 0x00000004) == 0x00000004)) { - switchTypes_ = java.util.Collections.unmodifiableList(switchTypes_); - bitField0_ = (bitField0_ & ~0x00000004); - } - result.switchTypes_ = switchTypes_; - result.bitField0_ = to_bitField0_; + public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest buildPartial() { + org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest(this); onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest)other); + if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest) { + return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest)other); } else { super.mergeFrom(other); return this; } } - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest.getDefaultInstance()) return this; - if (other.hasEnabled()) { - setEnabled(other.getEnabled()); - } - if (other.hasSynchronous()) { - setSynchronous(other.getSynchronous()); - } - if (!other.switchTypes_.isEmpty()) { - if (switchTypes_.isEmpty()) { - switchTypes_ = other.switchTypes_; - bitField0_ = (bitField0_ & ~0x00000004); - } else { - ensureSwitchTypesIsMutable(); - switchTypes_.addAll(other.switchTypes_); - } - onChanged(); - } + public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest other) { + if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { - if (!hasEnabled()) { - - return false; - } return true; } @@ -29394,11 +29077,11 @@ public final class MasterProtos { com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parsedMessage = null; + org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest) e.getUnfinishedMessage(); + parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeRequest) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { @@ -29407,193 +29090,50 @@ public final class MasterProtos { } return this; } - private int bitField0_; - - // required bool enabled = 1; - private boolean enabled_ ; - /** - * <code>required bool enabled = 1;</code> - */ - public boolean hasEnabled() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * <code>required bool enabled = 1;</code> - */ - public boolean getEnabled() { - return enabled_; - } - /** - * <code>required bool enabled = 1;</code> - */ - public Builder setEnabled(boolean value) { - bitField0_ |= 0x00000001; - enabled_ = value; - onChanged(); - return this; - } - /** - * <code>required bool enabled = 1;</code> - */ - public Builder clearEnabled() { - bitField0_ = (bitField0_ & ~0x00000001); - enabled_ = false; - onChanged(); - return this; - } - - // optional bool synchronous = 2; - private boolean synchronous_ ; - /** - * <code>optional bool synchronous = 2;</code> - */ - public boolean hasSynchronous() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * <code>optional bool synchronous = 2;</code> - */ - public boolean getSynchronous() { - return synchronous_; - } - /** - * <code>optional bool synchronous = 2;</code> - */ - public Builder setSynchronous(boolean value) { - bitField0_ |= 0x00000002; - synchronous_ = value; - onChanged(); - return this; - } - /** - * <code>optional bool synchronous = 2;</code> - */ - public Builder clearSynchronous() { - bitField0_ = (bitField0_ & ~0x00000002); - synchronous_ = false; - onChanged(); - return this; - } - - // repeated .hbase.pb.MasterSwitchType switch_types = 3; - private java.util.List<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType> switchTypes_ = - java.util.Collections.emptyList(); - private void ensureSwitchTypesIsMutable() { - if (!((bitField0_ & 0x00000004) == 0x00000004)) { - switchTypes_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType>(switchTypes_); - bitField0_ |= 0x00000004; - } - } - /** - * <code>repeated .hbase.pb.MasterSwitchType switch_types = 3;</code> - */ - public java.util.List<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType> getSwitchTypesList() { - return java.util.Collections.unmodifiableList(switchTypes_); - } - /** - * <code>repeated .hbase.pb.MasterSwitchType switch_types = 3;</code> - */ - public int getSwitchTypesCount() { - return switchTypes_.size(); - } - /** - * <code>repeated .hbase.pb.MasterSwitchType switch_types = 3;</code> - */ - public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType getSwitchTypes(int index) { - return switchTypes_.get(index); - } - /** - * <code>repeated .hbase.pb.MasterSwitchType switch_types = 3;</code> - */ - public Builder setSwitchTypes( - int index, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType value) { - if (value == null) { - throw new NullPointerException(); - } - ensureSwitchTypesIsMutable(); - switchTypes_.set(index, value); - onChanged(); - return this; - } - /** - * <code>repeated .hbase.pb.MasterSwitchType switch_types = 3;</code> - */ - public Builder addSwitchTypes(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType value) { - if (value == null) { - throw new NullPointerException(); - } - ensureSwitchTypesIsMutable(); - switchTypes_.add(value); - onChanged(); - return this; - } - /** - * <code>repeated .hbase.pb.MasterSwitchType switch_types = 3;</code> - */ - public Builder addAllSwitchTypes( - java.lang.Iterable<? extends org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType> values) { - ensureSwitchTypesIsMutable(); - super.addAll(values, switchTypes_); - onChanged(); - return this; - } - /** - * <code>repeated .hbase.pb.MasterSwitchType switch_types = 3;</code> - */ - public Builder clearSwitchTypes() { - switchTypes_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000004); - onChanged(); - return this; - } - // @@protoc_insertion_point(builder_scope:hbase.pb.SetSplitOrMergeEnabledRequest) + // @@protoc_insertion_point(builder_scope:hbase.pb.NormalizeRequest) } static { - defaultInstance = new SetSplitOrMergeEnabledRequest(true); + defaultInstance = new NormalizeRequest(true); defaultInstance.initFields(); } - // @@protoc_insertion_point(class_scope:hbase.pb.SetSplitOrMergeEnabledRequest) + // @@protoc_insertion_point(class_scope:hbase.pb.NormalizeRequest) } - public interface SetSplitOrMergeEnabledResponseOrBuilder + public interface NormalizeResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // repeated bool prev_value = 1; - /** - * <code>repeated bool prev_value = 1;</code> - */ - java.util.List<java.lang.Boolean> getPrevValueList(); + // required bool normalizer_ran = 1; /** - * <code>repeated bool prev_value = 1;</code> + * <code>required bool normalizer_ran = 1;</code> */ - int getPrevValueCount(); + boolean hasNormalizerRan(); /** - * <code>repeated bool prev_value = 1;</code> + * <code>required bool normalizer_ran = 1;</code> */ - boolean getPrevValue(int index); + boolean getNormalizerRan(); } /** - * Protobuf type {@code hbase.pb.SetSplitOrMergeEnabledResponse} + * Protobuf type {@code hbase.pb.NormalizeResponse} */ - public static final class SetSplitOrMergeEnabledResponse extends + public static final class NormalizeResponse extends com.google.protobuf.GeneratedMessage - implements SetSplitOrMergeEnabledResponseOrBuilder { - // Use SetSplitOrMergeEnabledResponse.newBuilder() to construct. - private SetSplitOrMergeEnabledResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + implements NormalizeResponseOrBuilder { + // Use NormalizeResponse.newBuilder() to construct. + private NormalizeResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private SetSplitOrMergeEnabledResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private NormalizeResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - private static final SetSplitOrMergeEnabledResponse defaultInstance; - public static SetSplitOrMergeEnabledResponse getDefaultInstance() { + private static final NormalizeResponse defaultInstance; + public static NormalizeResponse getDefaultInstance() { return defaultInstance; } - public SetSplitOrMergeEnabledResponse getDefaultInstanceForType() { + public NormalizeResponse getDefaultInstanceForType() { return defaultInstance; } @@ -29603,7 +29143,7 @@ public final class MasterProtos { getUnknownFields() { return this.unknownFields; } - private SetSplitOrMergeEnabledResponse( + private NormalizeResponse( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { @@ -29627,24 +29167,8 @@ public final class MasterProtos { break; } case 8: { - if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) { - prevValue_ = new java.util.ArrayList<java.lang.Boolean>(); - mutable_bitField0_ |= 0x00000001; - } - prevValue_.add(input.readBool()); - break; - } - case 10: { - int length = input.readRawVarint32(); - int limit = input.pushLimit(length); - if (!((mutable_bitField0_ & 0x00000001) == 0x00000001) && input.getBytesUntilLimit() > 0) { - prevValue_ = new java.util.ArrayList<java.lang.Boolean>(); - mutable_bitField0_ |= 0x00000001; - } - while (input.getBytesUntilLimit() > 0) { - prevValue_.add(input.readBool()); - } - input.popLimit(limit); + bitField0_ |= 0x00000001; + normalizerRan_ = input.readBool(); break; } } @@ -29655,71 +29179,66 @@ public final class MasterProtos { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { - if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) { - prevValue_ = java.util.Collections.unmodifiableList(prevValue_); - } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledResponse_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeResponse_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledResponse_fieldAccessorTable + return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.Builder.class); + org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse.Builder.class); } - public static com.google.protobuf.Parser<SetSplitOrMergeEnabledResponse> PARSER = - new com.google.protobuf.AbstractParser<SetSplitOrMergeEnabledResponse>() { - public SetSplitOrMergeEnabledResponse parsePartialFrom( + public static com.google.protobuf.Parser<NormalizeResponse> PARSER = + new com.google.protobuf.AbstractParser<NormalizeResponse>() { + public NormalizeResponse parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - return new SetSplitOrMergeEnabledResponse(input, extensionRegistry); + return new NormalizeResponse(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser<SetSplitOrMergeEnabledResponse> getParserForType() { + public com.google.protobuf.Parser<NormalizeResponse> getParserForType() { return PARSER; } - // repeated bool prev_value = 1; - public static final int PREV_VALUE_FIELD_NUMBER = 1; - private java.util.List<java.lang.Boolean> prevValue_; - /** - * <code>repeated bool prev_value = 1;</code> - */ - public java.util.List<java.lang.Boolean> - getPrevValueList() { - return prevValue_; - } + private int bitField0_; + // required bool normalizer_ran = 1; + public static final int NORMALIZER_RAN_FIELD_NUMBER = 1; + private boolean normalizerRan_; /** - * <code>repeated bool prev_value = 1;</code> + * <code>required bool normalizer_ran = 1;</code> */ - public int getPrevValueCount() { - return prevValue_.size(); + public boolean hasNormalizerRan() { + return ((bitField0_ & 0x00000001) == 0x00000001); } /** - * <code>repeated bool prev_value = 1;</code> + * <code>required bool normalizer_ran = 1;</code> */ - public boolean getPrevValue(int index) { - return prevValue_.get(index); + public boolean getNormalizerRan() { + return normalizerRan_; } private void initFields() { - prevValue_ = java.util.Collections.emptyList(); + normalizerRan_ = false; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; + if (!hasNormalizerRan()) { + memoizedIsInitialized = 0; + return false; + } memoizedIsInitialized = 1; return true; } @@ -29727,8 +29246,8 @@ public final class MasterProtos { public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); - for (int i = 0; i < prevValue_.size(); i++) { - output.writeBool(1, prevValue_.get(i)); + if (((bitField0_ & 0x00000001) == 0x00000001)) { + output.writeBool(1, normalizerRan_); } getUnknownFields().writeTo(output); } @@ -29739,11 +29258,9 @@ public final class MasterProtos { if (size != -1) return size; size = 0; - { - int dataSize = 0; - dataSize = 1 * getPrevValueList().size(); - size += dataSize; - size += 1 * getPrevValueList().size(); + if (((bitField0_ & 0x00000001) == 0x00000001)) { + size += com.google.protobuf.CodedOutputStream + .computeBoolSize(1, normalizerRan_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; @@ -29762,14 +29279,17 @@ public final class MasterProtos { if (obj == this) { return true; } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse)) { + if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse)) { return super.equals(obj); } - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse) obj; + org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse) obj; boolean result = true; - result = result && getPrevValueList() - .equals(other.getPrevValueList()); + result = result && (hasNormalizerRan() == other.hasNormalizerRan()); + if (hasNormalizerRan()) { + result = result && (getNormalizerRan() + == other.getNormalizerRan()); + } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; @@ -29783,62 +29303,62 @@ public final class MasterProtos { } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); - if (getPrevValueCount() > 0) { - hash = (37 * hash) + PREV_VALUE_FIELD_NUMBER; - hash = (53 * hash) + getPrevValueList().hashCode(); + if (hasNormalizerRan()) { + hash = (37 * hash) + NORMALIZER_RAN_FIELD_NUMBER; + hash = (53 * hash) + hashBoolean(getNormalizerRan()); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseFrom(byte[] data) + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseDelimitedFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseDelimitedFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -29847,7 +29367,7 @@ public final class MasterProtos { public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse prototype) { + public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @@ -29859,24 +29379,24 @@ public final class MasterProtos { return builder; } /** - * Protobuf type {@code hbase.pb.SetSplitOrMergeEnabledResponse} + * Protobuf type {@code hbase.pb.NormalizeResponse} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> - implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponseOrBuilder { + implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponseOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledResponse_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeResponse_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledResponse_fieldAccessorTable + return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.Builder.class); + org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse.Builder.class); } - // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.newBuilder() + // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } @@ -29896,7 +29416,7 @@ public final class MasterProtos { public Builder clear() { super.clear(); - prevValue_ = java.util.Collections.emptyList(); + normalizerRan_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @@ -29907,59 +29427,57 @@ public final class MasterProtos { public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledResponse_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeResponse_descriptor; } - public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.getDefaultInstance(); + public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse getDefaultInstanceForType() { + return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse.getDefaultInstance(); } - public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse build() { - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse result = buildPartial(); + public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse build() { + org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } - public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse(this); + public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse buildPartial() { + org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse(this); int from_bitField0_ = bitField0_; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - prevValue_ = java.util.Collections.unmodifiableList(prevValue_); - bitField0_ = (bitField0_ & ~0x00000001); + int to_bitField0_ = 0; + if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + to_bitField0_ |= 0x00000001; } - result.prevValue_ = prevValue_; + result.normalizerRan_ = normalizerRan_; + result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse)other); + if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse) { + return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse)other); } else { super.mergeFrom(other); return this; } } - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.getDefaultInstance()) return this; - if (!other.prevValue_.isEmpty()) { - if (prevValue_.isEmpty()) { - prevValue_ = other.prevValue_; - bitField0_ = (bitField0_ & ~0x00000001); - } else { - ensurePrevValueIsMutable(); - prevValue_.addAll(other.prevValue_); - } - onChanged(); + public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse other) { + if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse.getDefaultInstance()) return this; + if (other.hasNormalizerRan()) { + setNormalizerRan(other.getNormalizerRan()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { + if (!hasNormalizerRan()) { + + return false; + } return true; } @@ -29967,11 +29485,11 @@ public final class MasterProtos { com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parsedMessage = null; + org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse) e.getUnfinishedMessage(); + parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.NormalizeResponse) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { @@ -29982,115 +29500,82 @@ public final class MasterProtos { } private int bitField0_; - // repeated bool prev_value = 1; - private java.util.List<java.lang.Boolean> prevValue_ = java.util.Collections.emptyList(); - private void ensurePrevValueIsMutable() { - if (!((bitField0_ & 0x00000001) == 0x00000001)) { - prevValue_ = new java.util.ArrayList<java.lang.Boolean>(prevValue_); - bitField0_ |= 0x00000001; - } - } - /** - * <code>repeated bool prev_value = 1;</code> - */ - public java.util.List<java.lang.Boolean> - getPrevValueList() { - return java.util.Collections.unmodifiableList(prevValue_); - } - /** - * <code>repeated bool prev_value = 1;</code> - */ - public int getPrevValueCount() { - return prevValue_.size(); - } - /** - * <code>repeated bool prev_value = 1;</code> - */ - public boolean getPrevValue(int index) { - return prevValue_.get(index); - } + // required bool normalizer_ran = 1; + private boolean normalizerRan_ ; /** - * <code>repeated bool prev_value = 1;</code> + * <code>required bool normalizer_ran = 1;</code> */ - public Builder setPrevValue( - int index, boolean value) { - ensurePrevValueIsMutable(); - prevValue_.set(index, value); - onChanged(); - return this; + public boolean hasNormalizerRan() { + return ((bitField0_ & 0x00000001) == 0x00000001); } /** - * <code>repeated bool prev_value = 1;</code> + * <code>required bool normalizer_ran = 1;</code> */ - public Builder addPrevValue(boolean value) { - ensurePrevValueIsMutable(); - prevValue_.add(value); - onChanged(); - return this; + public boolean getNormalizerRan() { + return normalizerRan_; } /** - * <code>repeated bool prev_value = 1;</code> + * <code>required bool normalizer_ran = 1;</code> */ - public Builder addAllPrevValue( - java.lang.Iterable<? extends java.lang.Boolean> values) { - ensurePrevValueIsMutable(); - super.addAll(values, prevValue_); + public Builder setNormalizerRan(boolean value) { + bitField0_ |= 0x00000001; + normalizerRan_ = value; onChanged(); return this; } /** - * <code>repeated bool prev_value = 1;</code> + * <code>required bool normalizer_ran = 1;</code> */ - public Builder clearPrevValue() { - prevValue_ = java.util.Collections.emptyList(); + public Builder clearNormalizerRan() { bitField0_ = (bitField0_ & ~0x00000001); + normalizerRan_ = false; onChanged(); return this; } - // @@protoc_insertion_point(builder_scope:hbase.pb.SetSplitOrMergeEnabledResponse) + // @@protoc_insertion_point(builder_scope:hbase.pb.NormalizeResponse) } static { - defaultInstance = new SetSplitOrMergeEnabledResponse(true); + defaultInstance = new NormalizeResponse(true); defaultInstance.initFields(); } - // @@protoc_insertion_point(class_scope:hbase.pb.SetSplitOrMergeEnabledResponse) + // @@protoc_insertion_point(class_scope:hbase.pb.NormalizeResponse) } - public interface IsSplitOrMergeEnabledRequestOrBuilder + public interface SetNormalizerRunningRequestOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required .hbase.pb.MasterSwitchType switch_type = 1; + // required bool on = 1; /** - * <code>required .hbase.pb.MasterSwitchType switch_type = 1;</code> + * <code>required bool on = 1;</code> */ - boolean hasSwitchType(); + boolean hasOn(); /** - * <code>required .hbase.pb.MasterSwitchType switch_type = 1;</code> + * <code>required bool on = 1;</code> */ - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType getSwitchType(); + boolean getOn(); } /** - * Protobuf type {@code hbase.pb.IsSplitOrMergeEnabledRequest} + * Protobuf type {@code hbase.pb.SetNormalizerRunningRequest} */ - public static final class IsSplitOrMergeEnabledRequest extends + public static final class SetNormalizerRunningRequest extends com.google.protobuf.GeneratedMessage - implements IsSplitOrMergeEnabledRequestOrBuilder { - // Use IsSplitOrMergeEnabledRequest.newBuilder() to construct. - private IsSplitOrMergeEnabledRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + implements SetNormalizerRunningRequestOrBuilder { + // Use SetNormalizerRunningRequest.newBuilder() to construct. + private SetNormalizerRunningRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private IsSplitOrMergeEnabledRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private SetNormalizerRunningRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - private static final IsSplitOrMergeEnabledRequest defaultInstance; - public static IsSplitOrMergeEnabledRequest getDefaultInstance() { + private static final SetNormalizerRunningRequest defaultInstance; + public static SetNormalizerRunningRequest getDefaultInstance() { return defaultInstance; } - public IsSplitOrMergeEnabledRequest getDefaultInstanceForType() { + public SetNormalizerRunningRequest getDefaultInstanceForType() { return defaultInstance; } @@ -30100,7 +29585,7 @@ public final class MasterProtos { getUnknownFields() { return this.unknownFields; } - private IsSplitOrMergeEnabledRequest( + private SetNormalizerRunningRequest( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { @@ -30124,14 +29609,8 @@ public final class MasterProtos { break; } case 8: { - int rawValue = input.readEnum(); - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType value = org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType.valueOf(rawValue); - if (value == null) { - unknownFields.mergeVarintField(1, rawValue); - } else { - bitField0_ |= 0x00000001; - switchType_ = value; - } + bitField0_ |= 0x00000001; + on_ = input.readBool(); break; } } @@ -30148,57 +29627,57 @@ public final class MasterProtos { } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSplitOrMergeEnabledRequest_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetNormalizerRunningRequest_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSplitOrMergeEnabledRequest_fieldAccessorTable + return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetNormalizerRunningRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest.Builder.class); + org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest.Builder.class); } - public static com.google.protobuf.Parser<IsSplitOrMergeEnabledRequest> PARSER = - new com.google.protobuf.AbstractParser<IsSplitOrMergeEnabledRequest>() { - public IsSplitOrMergeEnabledRequest parsePartialFrom( + public static com.google.protobuf.Parser<SetNormalizerRunningRequest> PARSER = + new com.google.protobuf.AbstractParser<SetNormalizerRunningRequest>() { + public SetNormalizerRunningRequest parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - return new IsSplitOrMergeEnabledRequest(input, extensionRegistry); + return new SetNormalizerRunningRequest(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser<IsSplitOrMergeEnabledRequest> getParserForType() { + public com.google.protobuf.Parser<SetNormalizerRunningRequest> getParserForType() { return PARSER; } private int bitField0_; - // required .hbase.pb.MasterSwitchType switch_type = 1; - public static final int SWITCH_TYPE_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType switchType_; + // required bool on = 1; + public static final int ON_FIELD_NUMBER = 1; + private boolean on_; /** - * <code>required .hbase.pb.MasterSwitchType switch_type = 1;</code> + * <code>required bool on = 1;</code> */ - public boolean hasSwitchType() { + public boolean hasOn() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** - * <code>required .hbase.pb.MasterSwitchType switch_type = 1;</code> + * <code>required bool on = 1;</code> */ - public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType getSwitchType() { - return switchType_; + public boolean getOn() { + return on_; } private void initFields() { - switchType_ = org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType.SPLIT; + on_ = false; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; - if (!hasSwitchType()) { + if (!hasOn()) { memoizedIsInitialized = 0; return false; } @@ -30210,7 +29689,7 @@ public final class MasterProtos { throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeEnum(1, switchType_.getNumber()); + output.writeBool(1, on_); } getUnknownFields().writeTo(output); } @@ -30223,7 +29702,7 @@ public final class MasterProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream - .computeEnumSize(1, switchType_.getNumber()); + .computeBoolSize(1, on_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; @@ -30242,16 +29721,16 @@ public final class MasterProtos { if (obj == this) { return true; } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest)) { + if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest)) { return super.equals(obj); } - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest) obj; + org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest other = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest) obj; boolean result = true; - result = result && (hasSwitchType() == other.hasSwitchType()); - if (hasSwitchType()) { - result = result && - (getSwitchType() == other.getSwitchType()); + result = result && (hasOn() == other.hasOn()); + if (hasOn()) { + result = result && (getOn() + == other.getOn()); } result = result && getUnknownFields().equals(other.getUnknownFields()); @@ -30266,62 +29745,62 @@ public final class MasterProtos { } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasSwitchType()) { - hash = (37 * hash) + SWITCH_TYPE_FIELD_NUMBER; - hash = (53 * hash) + hashEnum(getSwitchType()); + if (hasOn()) { + hash = (37 * hash) + ON_FIELD_NUMBER; + hash = (53 * hash) + hashBoolean(getOn()); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseFrom(byte[] data) + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseDelimitedFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseDelimitedFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } - public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -30330,7 +29809,7 @@ public final class MasterProtos { public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest prototype) { + public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @@ -30342,24 +29821,24 @@ public final class MasterProtos { return builder; } /** - * Protobuf type {@code hbase.pb.IsSplitOrMergeEnabledRequest} + * Protobuf type {@code hbase.pb.SetNormalizerRunningRequest} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> - implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequestOrBuilder { + implements org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequestOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSplitOrMergeEnabledRequest_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetNormalizerRunningRequest_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSplitOrMergeEnabledRequest_fieldAccessorTable + return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetNormalizerRunningRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest.Builder.class); + org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest.class, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest.Builder.class); } - // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest.newBuilder() + // Construct using org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } @@ -30379,7 +29858,7 @@ public final class MasterProtos { public Builder clear() { super.clear(); - switchType_ = org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType.SPLIT; + on_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @@ -30390,54 +29869,54 @@ public final class MasterProtos { public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSplitOrMergeEnabledRequest_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetNormalizerRunningRequest_descriptor; } - public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest.getDefaultInstance(); + public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest getDefaultInstanceForType() { + return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest.getDefaultInstance(); } - public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest build() { - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest result = buildPartial(); + public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest build() { + org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } - public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest(this); + public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest buildPartial() { + org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest result = new org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } - result.switchType_ = switchType_; + result.on_ = on_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest)other); + if (other instanceof org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest) { + return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest)other); } else { super.mergeFrom(other); return this; } } - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest.getDefaultInstance()) return this; - if (other.hasSwitchType()) { - setSwitchType(other.getSwitchType()); + public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest other) { + if (other == org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest.getDefaultInstance()) return this; + if (other.hasOn()) { + setOn(other.getOn()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { - if (!hasSwitchType()) { + if (!hasOn()) { return false; } @@ -30448,11 +29927,11 @@ public final class MasterProtos { com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parsedMessage = null; + org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest) e.getUnfinishedMessage(); + parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SetNormalizerRunningRequest) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { @@ -30463,85 +29942,82 @@ public final class MasterProtos { } private int bitField0_; - // required .hbase.pb.MasterSwitchType switch_type = 1; - private org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType switchType_ = org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType.SPLIT; + // required bool on = 1; + private boolean on_ ; /** - * <code>required .hbase.pb.MasterSwitchType switch_type = 1;</code> + * <code>required bool on = 1;</code> */ - public boolean hasSwitchType() { + public boolean hasOn() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** - * <code>required .hbase.pb.MasterSwitchType switch_type = 1;</code> + * <code>required bool on = 1;</code> */ - public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType getSwitchType() { - return switchType_; + public boolean getOn() { + return on_; } /** - * <code>required .hbase.pb.MasterSwitchType switch_type = 1;</code> + * <code>required bool on = 1;</code> */ - public Builder setSwitchType(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType value) { - if (value == null) { - throw new NullPointerException(); - } + public Builder setOn(boolean value) { bitField0_ |= 0x00000001; - switchType_ = value; + on_ = value; onChanged(); return this; } /** - * <code>required .hbase.pb.MasterSwitchType switch_type = 1;</code> + * <code>required bool on = 1;</code> */ - public Builder clearSwitchType() { + public Builder clearOn() { bitField0_ = (bitField0_ & ~0x00000001); - switchType_ = org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterSwitchType.SPLIT; + on_ = false; onChanged(); return this; } - // @@protoc_insertion_point(builder_scope:hbase.pb.IsSplitOrMergeEnabledRequest) + // @@protoc_insertion_point(builder_scope:hbase.pb.SetNormalizerRunningRequest) } static { - defaultInstance = new IsSplitOrMergeEnabledRequest(true); + defaultInstance = new SetNormalizerRunningRequest(true); defaultInstance.initFields(); } - // @@protoc_insertion_point(class_scope:hbase.pb.IsSplitOrMergeEnabledRequest) + // @@protoc_insertion_point(class_scope:hbase.pb.SetNormalizerRunningRequest) } - public interface IsSplitOrMergeEnabledResponseOrBuilder + public interface SetNormalizerRunningResponseOrBuilder extends com.google.protobuf.MessageOrBuilder { - // required bool enabled = 1; + // optional bool prev_normalizer_value = 1; /** - * <code>required bool enabled = 1;</code> + * <code>optional bool prev_normalizer_value = 1;</code> */ - boolean hasEnabled(); + boolean hasPrevNormalizerValue(); /** - * <code>required bool enabled = 1;</code> + * <code>optional bool prev_normalizer_value = 1;</code> */ - boolean getEnabled(); + boolean getPrevNormalizerValue(); } /** - * Protobuf type {@code hbase.pb.IsSplitOrMergeEnabledResponse} + * Protobuf type {@code hbase.pb.SetNormalizerRunningResponse} */ - public static final class IsSplitOrMergeEnabledResponse extends + public static final class SetNormalizerRunningResponse extends com.google.protobuf.GeneratedMessage - implements IsSplitOrMergeEnabledResponseOrBuilder { - // Use IsSplitOrMergeEnabledResponse.newBuilder() to construct. - private IsSplitOrMergeEnabledResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + implements SetNormalizerRunningResponseOrBuilder { + // Use SetNormalizerRunningResponse.newBuilder() to construct. + private SetNormalizerRunningResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private IsSplitOrMergeEnabledResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private SetNormalizerRunningResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - private static final IsSplitOrMergeEnabledResponse defaultInstance; - public static IsSplitOrMergeEnabledResponse getDefaultInstance() { + private static final SetNormalizerRunningResponse defaultInstance; + public static SetNormalizerRunningResponse getDefaultInstance() { return defaultInstance; } - public IsSplitOrMergeEnabledResponse getDefaultInstanceForType() { + public SetNormalizerRunningRes
<TRUNCATED>