http://git-wip-us.apache.org/repos/asf/hbase/blob/df93c13f/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AggregateProtos.java ---------------------------------------------------------------------- diff --git a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AggregateProtos.java b/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AggregateProtos.java deleted file mode 100644 index 508790c..0000000 --- a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AggregateProtos.java +++ /dev/null @@ -1,2375 +0,0 @@ -// Generated by the protocol buffer compiler. DO NOT EDIT! -// source: Aggregate.proto - -package org.apache.hadoop.hbase.protobuf.generated; - -public final class AggregateProtos { - private AggregateProtos() {} - public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { - } - public interface AggregateRequestOrBuilder - extends com.google.protobuf.MessageOrBuilder { - - // required string interpreter_class_name = 1; - /** - * <code>required string interpreter_class_name = 1;</code> - * - * <pre> - ** The request passed to the AggregateService consists of three parts - * (1) the (canonical) classname of the ColumnInterpreter implementation - * (2) the Scan query - * (3) any bytes required to construct the ColumnInterpreter object - * properly - * </pre> - */ - boolean hasInterpreterClassName(); - /** - * <code>required string interpreter_class_name = 1;</code> - * - * <pre> - ** The request passed to the AggregateService consists of three parts - * (1) the (canonical) classname of the ColumnInterpreter implementation - * (2) the Scan query - * (3) any bytes required to construct the ColumnInterpreter object - * properly - * </pre> - */ - java.lang.String getInterpreterClassName(); - /** - * <code>required string interpreter_class_name = 1;</code> - * - * <pre> - ** The request passed to the AggregateService consists of three parts - * (1) the (canonical) classname of the ColumnInterpreter implementation - * (2) the Scan query - * (3) any bytes required to construct the ColumnInterpreter object - * properly - * </pre> - */ - com.google.protobuf.ByteString - getInterpreterClassNameBytes(); - - // required .hbase.pb.Scan scan = 2; - /** - * <code>required .hbase.pb.Scan scan = 2;</code> - */ - boolean hasScan(); - /** - * <code>required .hbase.pb.Scan scan = 2;</code> - */ - org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan getScan(); - /** - * <code>required .hbase.pb.Scan scan = 2;</code> - */ - org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder getScanOrBuilder(); - - // optional bytes interpreter_specific_bytes = 3; - /** - * <code>optional bytes interpreter_specific_bytes = 3;</code> - */ - boolean hasInterpreterSpecificBytes(); - /** - * <code>optional bytes interpreter_specific_bytes = 3;</code> - */ - com.google.protobuf.ByteString getInterpreterSpecificBytes(); - } - /** - * Protobuf type {@code hbase.pb.AggregateRequest} - */ - public static final class AggregateRequest extends - com.google.protobuf.GeneratedMessage - implements AggregateRequestOrBuilder { - // Use AggregateRequest.newBuilder() to construct. - private AggregateRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) { - super(builder); - this.unknownFields = builder.getUnknownFields(); - } - private AggregateRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final AggregateRequest defaultInstance; - public static AggregateRequest getDefaultInstance() { - return defaultInstance; - } - - public AggregateRequest getDefaultInstanceForType() { - return defaultInstance; - } - - private final com.google.protobuf.UnknownFieldSet unknownFields; - @java.lang.Override - public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { - return this.unknownFields; - } - private AggregateRequest( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); - int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); - try { - boolean done = false; - while (!done) { - int tag = input.readTag(); - switch (tag) { - case 0: - done = true; - break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } - case 10: { - bitField0_ |= 0x00000001; - interpreterClassName_ = input.readBytes(); - break; - } - case 18: { - org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder subBuilder = null; - if (((bitField0_ & 0x00000002) == 0x00000002)) { - subBuilder = scan_.toBuilder(); - } - scan_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.PARSER, extensionRegistry); - if (subBuilder != null) { - subBuilder.mergeFrom(scan_); - scan_ = subBuilder.buildPartial(); - } - bitField0_ |= 0x00000002; - break; - } - case 26: { - bitField0_ |= 0x00000004; - interpreterSpecificBytes_ = input.readBytes(); - break; - } - } - } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { - throw e.setUnfinishedMessage(this); - } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); - } finally { - this.unknownFields = unknownFields.build(); - makeExtensionsImmutable(); - } - } - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateRequest_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateRequest_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.class, org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.Builder.class); - } - - public static com.google.protobuf.Parser<AggregateRequest> PARSER = - new com.google.protobuf.AbstractParser<AggregateRequest>() { - public AggregateRequest parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new AggregateRequest(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser<AggregateRequest> getParserForType() { - return PARSER; - } - - private int bitField0_; - // required string interpreter_class_name = 1; - public static final int INTERPRETER_CLASS_NAME_FIELD_NUMBER = 1; - private java.lang.Object interpreterClassName_; - /** - * <code>required string interpreter_class_name = 1;</code> - * - * <pre> - ** The request passed to the AggregateService consists of three parts - * (1) the (canonical) classname of the ColumnInterpreter implementation - * (2) the Scan query - * (3) any bytes required to construct the ColumnInterpreter object - * properly - * </pre> - */ - public boolean hasInterpreterClassName() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * <code>required string interpreter_class_name = 1;</code> - * - * <pre> - ** The request passed to the AggregateService consists of three parts - * (1) the (canonical) classname of the ColumnInterpreter implementation - * (2) the Scan query - * (3) any bytes required to construct the ColumnInterpreter object - * properly - * </pre> - */ - public java.lang.String getInterpreterClassName() { - java.lang.Object ref = interpreterClassName_; - if (ref instanceof java.lang.String) { - return (java.lang.String) ref; - } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; - java.lang.String s = bs.toStringUtf8(); - if (bs.isValidUtf8()) { - interpreterClassName_ = s; - } - return s; - } - } - /** - * <code>required string interpreter_class_name = 1;</code> - * - * <pre> - ** The request passed to the AggregateService consists of three parts - * (1) the (canonical) classname of the ColumnInterpreter implementation - * (2) the Scan query - * (3) any bytes required to construct the ColumnInterpreter object - * properly - * </pre> - */ - public com.google.protobuf.ByteString - getInterpreterClassNameBytes() { - java.lang.Object ref = interpreterClassName_; - if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - interpreterClassName_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - - // required .hbase.pb.Scan scan = 2; - public static final int SCAN_FIELD_NUMBER = 2; - private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan scan_; - /** - * <code>required .hbase.pb.Scan scan = 2;</code> - */ - public boolean hasScan() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * <code>required .hbase.pb.Scan scan = 2;</code> - */ - public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan getScan() { - return scan_; - } - /** - * <code>required .hbase.pb.Scan scan = 2;</code> - */ - public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder getScanOrBuilder() { - return scan_; - } - - // optional bytes interpreter_specific_bytes = 3; - public static final int INTERPRETER_SPECIFIC_BYTES_FIELD_NUMBER = 3; - private com.google.protobuf.ByteString interpreterSpecificBytes_; - /** - * <code>optional bytes interpreter_specific_bytes = 3;</code> - */ - public boolean hasInterpreterSpecificBytes() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - /** - * <code>optional bytes interpreter_specific_bytes = 3;</code> - */ - public com.google.protobuf.ByteString getInterpreterSpecificBytes() { - return interpreterSpecificBytes_; - } - - private void initFields() { - interpreterClassName_ = ""; - scan_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.getDefaultInstance(); - interpreterSpecificBytes_ = com.google.protobuf.ByteString.EMPTY; - } - private byte memoizedIsInitialized = -1; - public final boolean isInitialized() { - byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; - - if (!hasInterpreterClassName()) { - memoizedIsInitialized = 0; - return false; - } - if (!hasScan()) { - memoizedIsInitialized = 0; - return false; - } - if (!getScan().isInitialized()) { - memoizedIsInitialized = 0; - return false; - } - memoizedIsInitialized = 1; - return true; - } - - public void writeTo(com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getInterpreterClassNameBytes()); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeMessage(2, scan_); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeBytes(3, interpreterSpecificBytes_); - } - getUnknownFields().writeTo(output); - } - - private int memoizedSerializedSize = -1; - public int getSerializedSize() { - int size = memoizedSerializedSize; - if (size != -1) return size; - - size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getInterpreterClassNameBytes()); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(2, scan_); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(3, interpreterSpecificBytes_); - } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; - return size; - } - - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override - public boolean equals(final java.lang.Object obj) { - if (obj == this) { - return true; - } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)) { - return super.equals(obj); - } - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest other = (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest) obj; - - boolean result = true; - result = result && (hasInterpreterClassName() == other.hasInterpreterClassName()); - if (hasInterpreterClassName()) { - result = result && getInterpreterClassName() - .equals(other.getInterpreterClassName()); - } - result = result && (hasScan() == other.hasScan()); - if (hasScan()) { - result = result && getScan() - .equals(other.getScan()); - } - result = result && (hasInterpreterSpecificBytes() == other.hasInterpreterSpecificBytes()); - if (hasInterpreterSpecificBytes()) { - result = result && getInterpreterSpecificBytes() - .equals(other.getInterpreterSpecificBytes()); - } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; - } - - private int memoizedHashCode = 0; - @java.lang.Override - public int hashCode() { - if (memoizedHashCode != 0) { - return memoizedHashCode; - } - int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasInterpreterClassName()) { - hash = (37 * hash) + INTERPRETER_CLASS_NAME_FIELD_NUMBER; - hash = (53 * hash) + getInterpreterClassName().hashCode(); - } - if (hasScan()) { - hash = (37 * hash) + SCAN_FIELD_NUMBER; - hash = (53 * hash) + getScan().hashCode(); - } - if (hasInterpreterSpecificBytes()) { - hash = (37 * hash) + INTERPRETER_SPECIFIC_BYTES_FIELD_NUMBER; - hash = (53 * hash) + getInterpreterSpecificBytes().hashCode(); - } - hash = (29 * hash) + getUnknownFields().hashCode(); - memoizedHashCode = hash; - return hash; - } - - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseFrom( - byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseFrom(java.io.InputStream input) - throws java.io.IOException { - return PARSER.parseFrom(input); - } - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); - } - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseDelimitedFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseFrom( - com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return PARSER.parseFrom(input); - } - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); - } - - public static Builder newBuilder() { return Builder.create(); } - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest prototype) { - return newBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { return newBuilder(this); } - - @java.lang.Override - protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - Builder builder = new Builder(parent); - return builder; - } - /** - * Protobuf type {@code hbase.pb.AggregateRequest} - */ - public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder<Builder> - implements org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequestOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateRequest_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateRequest_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.class, org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.Builder.class); - } - - // Construct using org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.newBuilder() - private Builder() { - maybeForceBuilderInitialization(); - } - - private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - super(parent); - maybeForceBuilderInitialization(); - } - private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { - getScanFieldBuilder(); - } - } - private static Builder create() { - return new Builder(); - } - - public Builder clear() { - super.clear(); - interpreterClassName_ = ""; - bitField0_ = (bitField0_ & ~0x00000001); - if (scanBuilder_ == null) { - scan_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.getDefaultInstance(); - } else { - scanBuilder_.clear(); - } - bitField0_ = (bitField0_ & ~0x00000002); - interpreterSpecificBytes_ = com.google.protobuf.ByteString.EMPTY; - bitField0_ = (bitField0_ & ~0x00000004); - return this; - } - - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - - public com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateRequest_descriptor; - } - - public org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance(); - } - - public org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest build() { - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException(result); - } - return result; - } - - public org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest result = new org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest(this); - int from_bitField0_ = bitField0_; - int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { - to_bitField0_ |= 0x00000001; - } - result.interpreterClassName_ = interpreterClassName_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { - to_bitField0_ |= 0x00000002; - } - if (scanBuilder_ == null) { - result.scan_ = scan_; - } else { - result.scan_ = scanBuilder_.build(); - } - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { - to_bitField0_ |= 0x00000004; - } - result.interpreterSpecificBytes_ = interpreterSpecificBytes_; - result.bitField0_ = to_bitField0_; - onBuilt(); - return result; - } - - public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance()) return this; - if (other.hasInterpreterClassName()) { - bitField0_ |= 0x00000001; - interpreterClassName_ = other.interpreterClassName_; - onChanged(); - } - if (other.hasScan()) { - mergeScan(other.getScan()); - } - if (other.hasInterpreterSpecificBytes()) { - setInterpreterSpecificBytes(other.getInterpreterSpecificBytes()); - } - this.mergeUnknownFields(other.getUnknownFields()); - return this; - } - - public final boolean isInitialized() { - if (!hasInterpreterClassName()) { - - return false; - } - if (!hasScan()) { - - return false; - } - if (!getScan().isInitialized()) { - - return false; - } - return true; - } - - public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parsedMessage = null; - try { - parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest) e.getUnfinishedMessage(); - throw e; - } finally { - if (parsedMessage != null) { - mergeFrom(parsedMessage); - } - } - return this; - } - private int bitField0_; - - // required string interpreter_class_name = 1; - private java.lang.Object interpreterClassName_ = ""; - /** - * <code>required string interpreter_class_name = 1;</code> - * - * <pre> - ** The request passed to the AggregateService consists of three parts - * (1) the (canonical) classname of the ColumnInterpreter implementation - * (2) the Scan query - * (3) any bytes required to construct the ColumnInterpreter object - * properly - * </pre> - */ - public boolean hasInterpreterClassName() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * <code>required string interpreter_class_name = 1;</code> - * - * <pre> - ** The request passed to the AggregateService consists of three parts - * (1) the (canonical) classname of the ColumnInterpreter implementation - * (2) the Scan query - * (3) any bytes required to construct the ColumnInterpreter object - * properly - * </pre> - */ - public java.lang.String getInterpreterClassName() { - java.lang.Object ref = interpreterClassName_; - if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - interpreterClassName_ = s; - return s; - } else { - return (java.lang.String) ref; - } - } - /** - * <code>required string interpreter_class_name = 1;</code> - * - * <pre> - ** The request passed to the AggregateService consists of three parts - * (1) the (canonical) classname of the ColumnInterpreter implementation - * (2) the Scan query - * (3) any bytes required to construct the ColumnInterpreter object - * properly - * </pre> - */ - public com.google.protobuf.ByteString - getInterpreterClassNameBytes() { - java.lang.Object ref = interpreterClassName_; - if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - interpreterClassName_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - /** - * <code>required string interpreter_class_name = 1;</code> - * - * <pre> - ** The request passed to the AggregateService consists of three parts - * (1) the (canonical) classname of the ColumnInterpreter implementation - * (2) the Scan query - * (3) any bytes required to construct the ColumnInterpreter object - * properly - * </pre> - */ - public Builder setInterpreterClassName( - java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000001; - interpreterClassName_ = value; - onChanged(); - return this; - } - /** - * <code>required string interpreter_class_name = 1;</code> - * - * <pre> - ** The request passed to the AggregateService consists of three parts - * (1) the (canonical) classname of the ColumnInterpreter implementation - * (2) the Scan query - * (3) any bytes required to construct the ColumnInterpreter object - * properly - * </pre> - */ - public Builder clearInterpreterClassName() { - bitField0_ = (bitField0_ & ~0x00000001); - interpreterClassName_ = getDefaultInstance().getInterpreterClassName(); - onChanged(); - return this; - } - /** - * <code>required string interpreter_class_name = 1;</code> - * - * <pre> - ** The request passed to the AggregateService consists of three parts - * (1) the (canonical) classname of the ColumnInterpreter implementation - * (2) the Scan query - * (3) any bytes required to construct the ColumnInterpreter object - * properly - * </pre> - */ - public Builder setInterpreterClassNameBytes( - com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000001; - interpreterClassName_ = value; - onChanged(); - return this; - } - - // required .hbase.pb.Scan scan = 2; - private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan scan_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder> scanBuilder_; - /** - * <code>required .hbase.pb.Scan scan = 2;</code> - */ - public boolean hasScan() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * <code>required .hbase.pb.Scan scan = 2;</code> - */ - public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan getScan() { - if (scanBuilder_ == null) { - return scan_; - } else { - return scanBuilder_.getMessage(); - } - } - /** - * <code>required .hbase.pb.Scan scan = 2;</code> - */ - public Builder setScan(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan value) { - if (scanBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - scan_ = value; - onChanged(); - } else { - scanBuilder_.setMessage(value); - } - bitField0_ |= 0x00000002; - return this; - } - /** - * <code>required .hbase.pb.Scan scan = 2;</code> - */ - public Builder setScan( - org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder builderForValue) { - if (scanBuilder_ == null) { - scan_ = builderForValue.build(); - onChanged(); - } else { - scanBuilder_.setMessage(builderForValue.build()); - } - bitField0_ |= 0x00000002; - return this; - } - /** - * <code>required .hbase.pb.Scan scan = 2;</code> - */ - public Builder mergeScan(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan value) { - if (scanBuilder_ == null) { - if (((bitField0_ & 0x00000002) == 0x00000002) && - scan_ != org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.getDefaultInstance()) { - scan_ = - org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.newBuilder(scan_).mergeFrom(value).buildPartial(); - } else { - scan_ = value; - } - onChanged(); - } else { - scanBuilder_.mergeFrom(value); - } - bitField0_ |= 0x00000002; - return this; - } - /** - * <code>required .hbase.pb.Scan scan = 2;</code> - */ - public Builder clearScan() { - if (scanBuilder_ == null) { - scan_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.getDefaultInstance(); - onChanged(); - } else { - scanBuilder_.clear(); - } - bitField0_ = (bitField0_ & ~0x00000002); - return this; - } - /** - * <code>required .hbase.pb.Scan scan = 2;</code> - */ - public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder getScanBuilder() { - bitField0_ |= 0x00000002; - onChanged(); - return getScanFieldBuilder().getBuilder(); - } - /** - * <code>required .hbase.pb.Scan scan = 2;</code> - */ - public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder getScanOrBuilder() { - if (scanBuilder_ != null) { - return scanBuilder_.getMessageOrBuilder(); - } else { - return scan_; - } - } - /** - * <code>required .hbase.pb.Scan scan = 2;</code> - */ - private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder> - getScanFieldBuilder() { - if (scanBuilder_ == null) { - scanBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder>( - scan_, - getParentForChildren(), - isClean()); - scan_ = null; - } - return scanBuilder_; - } - - // optional bytes interpreter_specific_bytes = 3; - private com.google.protobuf.ByteString interpreterSpecificBytes_ = com.google.protobuf.ByteString.EMPTY; - /** - * <code>optional bytes interpreter_specific_bytes = 3;</code> - */ - public boolean hasInterpreterSpecificBytes() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - /** - * <code>optional bytes interpreter_specific_bytes = 3;</code> - */ - public com.google.protobuf.ByteString getInterpreterSpecificBytes() { - return interpreterSpecificBytes_; - } - /** - * <code>optional bytes interpreter_specific_bytes = 3;</code> - */ - public Builder setInterpreterSpecificBytes(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000004; - interpreterSpecificBytes_ = value; - onChanged(); - return this; - } - /** - * <code>optional bytes interpreter_specific_bytes = 3;</code> - */ - public Builder clearInterpreterSpecificBytes() { - bitField0_ = (bitField0_ & ~0x00000004); - interpreterSpecificBytes_ = getDefaultInstance().getInterpreterSpecificBytes(); - onChanged(); - return this; - } - - // @@protoc_insertion_point(builder_scope:hbase.pb.AggregateRequest) - } - - static { - defaultInstance = new AggregateRequest(true); - defaultInstance.initFields(); - } - - // @@protoc_insertion_point(class_scope:hbase.pb.AggregateRequest) - } - - public interface AggregateResponseOrBuilder - extends com.google.protobuf.MessageOrBuilder { - - // repeated bytes first_part = 1; - /** - * <code>repeated bytes first_part = 1;</code> - * - * <pre> - ** - * The AggregateService methods all have a response that either is a Pair - * or a simple object. When it is a Pair both first_part and second_part - * have defined values (and the second_part is not present in the response - * when the response is not a pair). Refer to the AggregateImplementation - * class for an overview of the AggregateResponse object constructions. - * </pre> - */ - java.util.List<com.google.protobuf.ByteString> getFirstPartList(); - /** - * <code>repeated bytes first_part = 1;</code> - * - * <pre> - ** - * The AggregateService methods all have a response that either is a Pair - * or a simple object. When it is a Pair both first_part and second_part - * have defined values (and the second_part is not present in the response - * when the response is not a pair). Refer to the AggregateImplementation - * class for an overview of the AggregateResponse object constructions. - * </pre> - */ - int getFirstPartCount(); - /** - * <code>repeated bytes first_part = 1;</code> - * - * <pre> - ** - * The AggregateService methods all have a response that either is a Pair - * or a simple object. When it is a Pair both first_part and second_part - * have defined values (and the second_part is not present in the response - * when the response is not a pair). Refer to the AggregateImplementation - * class for an overview of the AggregateResponse object constructions. - * </pre> - */ - com.google.protobuf.ByteString getFirstPart(int index); - - // optional bytes second_part = 2; - /** - * <code>optional bytes second_part = 2;</code> - */ - boolean hasSecondPart(); - /** - * <code>optional bytes second_part = 2;</code> - */ - com.google.protobuf.ByteString getSecondPart(); - } - /** - * Protobuf type {@code hbase.pb.AggregateResponse} - */ - public static final class AggregateResponse extends - com.google.protobuf.GeneratedMessage - implements AggregateResponseOrBuilder { - // Use AggregateResponse.newBuilder() to construct. - private AggregateResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) { - super(builder); - this.unknownFields = builder.getUnknownFields(); - } - private AggregateResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final AggregateResponse defaultInstance; - public static AggregateResponse getDefaultInstance() { - return defaultInstance; - } - - public AggregateResponse getDefaultInstanceForType() { - return defaultInstance; - } - - private final com.google.protobuf.UnknownFieldSet unknownFields; - @java.lang.Override - public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { - return this.unknownFields; - } - private AggregateResponse( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); - int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); - try { - boolean done = false; - while (!done) { - int tag = input.readTag(); - switch (tag) { - case 0: - done = true; - break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } - case 10: { - if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) { - firstPart_ = new java.util.ArrayList<com.google.protobuf.ByteString>(); - mutable_bitField0_ |= 0x00000001; - } - firstPart_.add(input.readBytes()); - break; - } - case 18: { - bitField0_ |= 0x00000001; - secondPart_ = input.readBytes(); - break; - } - } - } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { - throw e.setUnfinishedMessage(this); - } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); - } finally { - if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) { - firstPart_ = java.util.Collections.unmodifiableList(firstPart_); - } - this.unknownFields = unknownFields.build(); - makeExtensionsImmutable(); - } - } - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateResponse_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateResponse_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.class, org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.Builder.class); - } - - public static com.google.protobuf.Parser<AggregateResponse> PARSER = - new com.google.protobuf.AbstractParser<AggregateResponse>() { - public AggregateResponse parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new AggregateResponse(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser<AggregateResponse> getParserForType() { - return PARSER; - } - - private int bitField0_; - // repeated bytes first_part = 1; - public static final int FIRST_PART_FIELD_NUMBER = 1; - private java.util.List<com.google.protobuf.ByteString> firstPart_; - /** - * <code>repeated bytes first_part = 1;</code> - * - * <pre> - ** - * The AggregateService methods all have a response that either is a Pair - * or a simple object. When it is a Pair both first_part and second_part - * have defined values (and the second_part is not present in the response - * when the response is not a pair). Refer to the AggregateImplementation - * class for an overview of the AggregateResponse object constructions. - * </pre> - */ - public java.util.List<com.google.protobuf.ByteString> - getFirstPartList() { - return firstPart_; - } - /** - * <code>repeated bytes first_part = 1;</code> - * - * <pre> - ** - * The AggregateService methods all have a response that either is a Pair - * or a simple object. When it is a Pair both first_part and second_part - * have defined values (and the second_part is not present in the response - * when the response is not a pair). Refer to the AggregateImplementation - * class for an overview of the AggregateResponse object constructions. - * </pre> - */ - public int getFirstPartCount() { - return firstPart_.size(); - } - /** - * <code>repeated bytes first_part = 1;</code> - * - * <pre> - ** - * The AggregateService methods all have a response that either is a Pair - * or a simple object. When it is a Pair both first_part and second_part - * have defined values (and the second_part is not present in the response - * when the response is not a pair). Refer to the AggregateImplementation - * class for an overview of the AggregateResponse object constructions. - * </pre> - */ - public com.google.protobuf.ByteString getFirstPart(int index) { - return firstPart_.get(index); - } - - // optional bytes second_part = 2; - public static final int SECOND_PART_FIELD_NUMBER = 2; - private com.google.protobuf.ByteString secondPart_; - /** - * <code>optional bytes second_part = 2;</code> - */ - public boolean hasSecondPart() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * <code>optional bytes second_part = 2;</code> - */ - public com.google.protobuf.ByteString getSecondPart() { - return secondPart_; - } - - private void initFields() { - firstPart_ = java.util.Collections.emptyList(); - secondPart_ = com.google.protobuf.ByteString.EMPTY; - } - private byte memoizedIsInitialized = -1; - public final boolean isInitialized() { - byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; - - memoizedIsInitialized = 1; - return true; - } - - public void writeTo(com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - getSerializedSize(); - for (int i = 0; i < firstPart_.size(); i++) { - output.writeBytes(1, firstPart_.get(i)); - } - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(2, secondPart_); - } - getUnknownFields().writeTo(output); - } - - private int memoizedSerializedSize = -1; - public int getSerializedSize() { - int size = memoizedSerializedSize; - if (size != -1) return size; - - size = 0; - { - int dataSize = 0; - for (int i = 0; i < firstPart_.size(); i++) { - dataSize += com.google.protobuf.CodedOutputStream - .computeBytesSizeNoTag(firstPart_.get(i)); - } - size += dataSize; - size += 1 * getFirstPartList().size(); - } - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, secondPart_); - } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; - return size; - } - - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override - public boolean equals(final java.lang.Object obj) { - if (obj == this) { - return true; - } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse)) { - return super.equals(obj); - } - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse other = (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse) obj; - - boolean result = true; - result = result && getFirstPartList() - .equals(other.getFirstPartList()); - result = result && (hasSecondPart() == other.hasSecondPart()); - if (hasSecondPart()) { - result = result && getSecondPart() - .equals(other.getSecondPart()); - } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; - } - - private int memoizedHashCode = 0; - @java.lang.Override - public int hashCode() { - if (memoizedHashCode != 0) { - return memoizedHashCode; - } - int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - if (getFirstPartCount() > 0) { - hash = (37 * hash) + FIRST_PART_FIELD_NUMBER; - hash = (53 * hash) + getFirstPartList().hashCode(); - } - if (hasSecondPart()) { - hash = (37 * hash) + SECOND_PART_FIELD_NUMBER; - hash = (53 * hash) + getSecondPart().hashCode(); - } - hash = (29 * hash) + getUnknownFields().hashCode(); - memoizedHashCode = hash; - return hash; - } - - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseFrom( - byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseFrom(java.io.InputStream input) - throws java.io.IOException { - return PARSER.parseFrom(input); - } - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); - } - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseDelimitedFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseFrom( - com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return PARSER.parseFrom(input); - } - public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); - } - - public static Builder newBuilder() { return Builder.create(); } - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse prototype) { - return newBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { return newBuilder(this); } - - @java.lang.Override - protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - Builder builder = new Builder(parent); - return builder; - } - /** - * Protobuf type {@code hbase.pb.AggregateResponse} - */ - public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder<Builder> - implements org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponseOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateResponse_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateResponse_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.class, org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.Builder.class); - } - - // Construct using org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.newBuilder() - private Builder() { - maybeForceBuilderInitialization(); - } - - private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - super(parent); - maybeForceBuilderInitialization(); - } - private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { - } - } - private static Builder create() { - return new Builder(); - } - - public Builder clear() { - super.clear(); - firstPart_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000001); - secondPart_ = com.google.protobuf.ByteString.EMPTY; - bitField0_ = (bitField0_ & ~0x00000002); - return this; - } - - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - - public com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateResponse_descriptor; - } - - public org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(); - } - - public org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse build() { - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException(result); - } - return result; - } - - public org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse result = new org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse(this); - int from_bitField0_ = bitField0_; - int to_bitField0_ = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - firstPart_ = java.util.Collections.unmodifiableList(firstPart_); - bitField0_ = (bitField0_ & ~0x00000001); - } - result.firstPart_ = firstPart_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { - to_bitField0_ |= 0x00000001; - } - result.secondPart_ = secondPart_; - result.bitField0_ = to_bitField0_; - onBuilt(); - return result; - } - - public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance()) return this; - if (!other.firstPart_.isEmpty()) { - if (firstPart_.isEmpty()) { - firstPart_ = other.firstPart_; - bitField0_ = (bitField0_ & ~0x00000001); - } else { - ensureFirstPartIsMutable(); - firstPart_.addAll(other.firstPart_); - } - onChanged(); - } - if (other.hasSecondPart()) { - setSecondPart(other.getSecondPart()); - } - this.mergeUnknownFields(other.getUnknownFields()); - return this; - } - - public final boolean isInitialized() { - return true; - } - - public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parsedMessage = null; - try { - parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse) e.getUnfinishedMessage(); - throw e; - } finally { - if (parsedMessage != null) { - mergeFrom(parsedMessage); - } - } - return this; - } - private int bitField0_; - - // repeated bytes first_part = 1; - private java.util.List<com.google.protobuf.ByteString> firstPart_ = java.util.Collections.emptyList(); - private void ensureFirstPartIsMutable() { - if (!((bitField0_ & 0x00000001) == 0x00000001)) { - firstPart_ = new java.util.ArrayList<com.google.protobuf.ByteString>(firstPart_); - bitField0_ |= 0x00000001; - } - } - /** - * <code>repeated bytes first_part = 1;</code> - * - * <pre> - ** - * The AggregateService methods all have a response that either is a Pair - * or a simple object. When it is a Pair both first_part and second_part - * have defined values (and the second_part is not present in the response - * when the response is not a pair). Refer to the AggregateImplementation - * class for an overview of the AggregateResponse object constructions. - * </pre> - */ - public java.util.List<com.google.protobuf.ByteString> - getFirstPartList() { - return java.util.Collections.unmodifiableList(firstPart_); - } - /** - * <code>repeated bytes first_part = 1;</code> - * - * <pre> - ** - * The AggregateService methods all have a response that either is a Pair - * or a simple object. When it is a Pair both first_part and second_part - * have defined values (and the second_part is not present in the response - * when the response is not a pair). Refer to the AggregateImplementation - * class for an overview of the AggregateResponse object constructions. - * </pre> - */ - public int getFirstPartCount() { - return firstPart_.size(); - } - /** - * <code>repeated bytes first_part = 1;</code> - * - * <pre> - ** - * The AggregateService methods all have a response that either is a Pair - * or a simple object. When it is a Pair both first_part and second_part - * have defined values (and the second_part is not present in the response - * when the response is not a pair). Refer to the AggregateImplementation - * class for an overview of the AggregateResponse object constructions. - * </pre> - */ - public com.google.protobuf.ByteString getFirstPart(int index) { - return firstPart_.get(index); - } - /** - * <code>repeated bytes first_part = 1;</code> - * - * <pre> - ** - * The AggregateService methods all have a response that either is a Pair - * or a simple object. When it is a Pair both first_part and second_part - * have defined values (and the second_part is not present in the response - * when the response is not a pair). Refer to the AggregateImplementation - * class for an overview of the AggregateResponse object constructions. - * </pre> - */ - public Builder setFirstPart( - int index, com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - ensureFirstPartIsMutable(); - firstPart_.set(index, value); - onChanged(); - return this; - } - /** - * <code>repeated bytes first_part = 1;</code> - * - * <pre> - ** - * The AggregateService methods all have a response that either is a Pair - * or a simple object. When it is a Pair both first_part and second_part - * have defined values (and the second_part is not present in the response - * when the response is not a pair). Refer to the AggregateImplementation - * class for an overview of the AggregateResponse object constructions. - * </pre> - */ - public Builder addFirstPart(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - ensureFirstPartIsMutable(); - firstPart_.add(value); - onChanged(); - return this; - } - /** - * <code>repeated bytes first_part = 1;</code> - * - * <pre> - ** - * The AggregateService methods all have a response that either is a Pair - * or a simple object. When it is a Pair both first_part and second_part - * have defined values (and the second_part is not present in the response - * when the response is not a pair). Refer to the AggregateImplementation - * class for an overview of the AggregateResponse object constructions. - * </pre> - */ - public Builder addAllFirstPart( - java.lang.Iterable<? extends com.google.protobuf.ByteString> values) { - ensureFirstPartIsMutable(); - super.addAll(values, firstPart_); - onChanged(); - return this; - } - /** - * <code>repeated bytes first_part = 1;</code> - * - * <pre> - ** - * The AggregateService methods all have a response that either is a Pair - * or a simple object. When it is a Pair both first_part and second_part - * have defined values (and the second_part is not present in the response - * when the response is not a pair). Refer to the AggregateImplementation - * class for an overview of the AggregateResponse object constructions. - * </pre> - */ - public Builder clearFirstPart() { - firstPart_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000001); - onChanged(); - return this; - } - - // optional bytes second_part = 2; - private com.google.protobuf.ByteString secondPart_ = com.google.protobuf.ByteString.EMPTY; - /** - * <code>optional bytes second_part = 2;</code> - */ - public boolean hasSecondPart() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * <code>optional bytes second_part = 2;</code> - */ - public com.google.protobuf.ByteString getSecondPart() { - return secondPart_; - } - /** - * <code>optional bytes second_part = 2;</code> - */ - public Builder setSecondPart(com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000002; - secondPart_ = value; - onChanged(); - return this; - } - /** - * <code>optional bytes second_part = 2;</code> - */ - public Builder clearSecondPart() { - bitField0_ = (bitField0_ & ~0x00000002); - secondPart_ = getDefaultInstance().getSecondPart(); - onChanged(); - return this; - } - - // @@protoc_insertion_point(builder_scope:hbase.pb.AggregateResponse) - } - - static { - defaultInstance = new AggregateResponse(true); - defaultInstance.initFields(); - } - - // @@protoc_insertion_point(class_scope:hbase.pb.AggregateResponse) - } - - /** - * Protobuf service {@code hbase.pb.AggregateService} - * - * <pre> - ** Refer to the AggregateImplementation class for an overview of the - * AggregateService method implementations and their functionality. - * </pre> - */ - public static abstract class AggregateService - implements com.google.protobuf.Service { - protected AggregateService() {} - - public interface Interface { - /** - * <code>rpc GetMax(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code> - */ - public abstract void getMax( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done); - - /** - * <code>rpc GetMin(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code> - */ - public abstract void getMin( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done); - - /** - * <code>rpc GetSum(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code> - */ - public abstract void getSum( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done); - - /** - * <code>rpc GetRowNum(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code> - */ - public abstract void getRowNum( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done); - - /** - * <code>rpc GetAvg(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code> - */ - public abstract void getAvg( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done); - - /** - * <code>rpc GetStd(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code> - */ - public abstract void getStd( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done); - - /** - * <code>rpc GetMedian(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code> - */ - public abstract void getMedian( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done); - - } - - public static com.google.protobuf.Service newReflectiveService( - final Interface impl) { - return new AggregateService() { - @java.lang.Override - public void getMax( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) { - impl.getMax(controller, request, done); - } - - @java.lang.Override - public void getMin( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) { - impl.getMin(controller, request, done); - } - - @java.lang.Override - public void getSum( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) { - impl.getSum(controller, request, done); - } - - @java.lang.Override - public void getRowNum( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) { - impl.getRowNum(controller, request, done); - } - - @java.lang.Override - public void getAvg( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) { - impl.getAvg(controller, request, done); - } - - @java.lang.Override - public void getStd( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) { - impl.getStd(controller, request, done); - } - - @java.lang.Override - public void getMedian( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) { - impl.getMedian(controller, request, done); - } - - }; - } - - public static com.google.protobuf.BlockingService - newReflectiveBlockingService(final BlockingInterface impl) { - return new com.google.protobuf.BlockingService() { - public final com.google.protobuf.Descriptors.ServiceDescriptor - getDescriptorForType() { - return getDescriptor(); - } - - public final com.google.protobuf.Message callBlockingMethod( - com.google.protobuf.Descriptors.MethodDescriptor method, - com.google.protobuf.RpcController controller, - com.google.protobuf.Message request) - throws com.google.protobuf.ServiceException { - if (method.getService() != getDescriptor()) { - throw new java.lang.IllegalArgumentException( - "Service.callBlockingMethod() given method descriptor for " + - "wrong service type."); - } - switch(method.getIndex()) { - case 0: - return impl.getMax(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request); - case 1: - return impl.getMin(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request); - case 2: - return impl.getSum(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request); - case 3: - return impl.getRowNum(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request); - case 4: - return impl.getAvg(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request); - case 5: - return impl.getStd(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request); - case 6: - return impl.getMedian(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request); - default: - throw new java.lang.AssertionError("Can't get here."); - } - } - - public final com.google.protobuf.Message - getRequestPrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { - if (method.getService() != getDescriptor()) { - throw new java.lang.IllegalArgumentException( - "Service.getRequestPrototype() given method " + - "descriptor for wrong service type."); - } - switch(method.getIndex()) { - case 0: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance(); - case 1: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance(); - case 2: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance(); - case 3: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance(); - case 4: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance(); - case 5: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance(); - case 6: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance(); - default: - throw new java.lang.AssertionError("Can't get here."); - } - } - - public final com.google.protobuf.Message - getResponsePrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { - if (method.getService() != getDescriptor()) { - throw new java.lang.IllegalArgumentException( - "Service.getResponsePrototype() given method " + - "descriptor for wrong service type."); - } - switch(method.getIndex()) { - case 0: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(); - case 1: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(); - case 2: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(); - case 3: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(); - case 4: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(); - case 5: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(); - case 6: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(); - default: - throw new java.lang.AssertionError("Can't get here."); - } - } - - }; - } - - /** - * <code>rpc GetMax(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code> - */ - public abstract void getMax( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done); - - /** - * <code>rpc GetMin(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code> - */ - public abstract void getMin( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done); - - /** - * <code>rpc GetSum(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code> - */ - public abstract void getSum( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done); - - /** - * <code>rpc GetRowNum(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code> - */ - public abstract void getRowNum( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done); - - /** - * <code>rpc GetAvg(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code> - */ - public abstract void getAvg( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done); - - /** - * <code>rpc GetStd(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code> - */ - public abstract void getStd( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done); - - /** - * <code>rpc GetMedian(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code> - */ - public abstract void getMedian( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done); - - public static final - com.google.protobuf.Descriptors.ServiceDescriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.getDescriptor().getServices().get(0); - } - public final com.google.protobuf.Descriptors.ServiceDescriptor - getDescriptorForType() { - return getDescriptor(); - } - - public final void callMethod( - com.google.protobuf.Descriptors.MethodDescriptor method, - com.google.protobuf.RpcController controller, - com.google.protobuf.Message request, - com.google.protobuf.RpcCallback< - com.google.protobuf.Message> done) { - if (method.getService() != getDescriptor()) { - throw new java.lang.IllegalArgumentException( - "Service.callMethod() given method descriptor for wrong " + - "service type."); - } - switch(method.getIndex()) { - case 0: - this.getMax(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request, - com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse>specializeCallback( - done)); - return; - case 1: - this.getMin(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request, - com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse>specializeCallback( - done)); - return; - case 2: - this.getSum(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request, - com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse>specializeCallback( - done)); - return; - case 3: - this.getRowNum(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request, - com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse>specializeCallback( - done)); - return; - case 4: - this.getAvg(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request, - com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse>specializeCallback( - done)); - return; - case 5: - this.getStd(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request, - com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse>specializeCallback( - done)); - return; - case 6: - this.getMedian(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request, - com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse>specializeCallback( - done)); - return; - default: - throw new java.lang.AssertionError("Can't get here."); - } - } - - public final com.google.protobuf.Message - getRequestPrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { - if (method.getService() != getDescriptor()) { - throw new java.lang.IllegalArgumentException( - "Service.getRequestPrototype() given method " + - "descriptor for wrong service type."); - } - switch(method.getIndex()) { - case 0: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance(); - case 1: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance(); - case 2: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance(); - case 3: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance(); - case 4: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance(); - case 5: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance(); - case 6: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance(); - default: - throw new java.lang.AssertionError("Can't get here."); - } - } - - public final com.google.protobuf.Message - getResponsePrototype( - com.google.protobuf.Descriptors.MethodDescriptor method) { - if (method.getService() != getDescriptor()) { - throw new java.lang.IllegalArgumentException( - "Service.getResponsePrototype() given method " + - "descriptor for wrong service type."); - } - switch(method.getIndex()) { - case 0: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(); - case 1: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(); - case 2: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(); - case 3: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(); - case 4: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(); - case 5: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(); - case 6: - return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(); - default: - throw new java.lang.AssertionError("Can't get here."); - } - } - - public static Stub newStub( - com.google.protobuf.RpcChannel channel) { - return new Stub(channel); - } - - public static final class Stub extends org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateService implements Interface { - private Stub(com.google.protobuf.RpcChannel channel) { - this.channel = channel; - } - - private final com.google.protobuf.RpcChannel channel; - - public com.google.protobuf.RpcChannel getChannel() { - return channel; - } - - public void getMax( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) { - channel.callMethod( - getDescriptor().getMethods().get(0), - controller, - request, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( - done, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.class, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance())); - } - - public void getMin( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) { - channel.callMethod( - getDescriptor().getMethods().get(1), - controller, - request, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( - done, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.class, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance())); - } - - public void getSum( - com.google.protobuf.RpcController controller, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request, - com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) { - channel.callMethod( - getDescriptor().getMethods().get(2), - controller, - request, - org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(), - com.google.protobuf.RpcUtil.generalizeCallback( - done, - org.apache.hadoop.hbase.
<TRUNCATED>