http://git-wip-us.apache.org/repos/asf/hbase/blob/c61f7c34/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/BackupProtos.java ---------------------------------------------------------------------- diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/BackupProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/BackupProtos.java index a1a1a78..4699c81 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/BackupProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/BackupProtos.java @@ -4412,85 +4412,55 @@ public final class BackupProtos { */ long getCompleteTs(); - // required int64 total_bytes = 7; + // repeated .hbase.pb.TableServerTimestamp tst_map = 7; /** - * <code>required int64 total_bytes = 7;</code> - */ - boolean hasTotalBytes(); - /** - * <code>required int64 total_bytes = 7;</code> - */ - long getTotalBytes(); - - // optional int64 log_bytes = 8; - /** - * <code>optional int64 log_bytes = 8;</code> - */ - boolean hasLogBytes(); - /** - * <code>optional int64 log_bytes = 8;</code> - */ - long getLogBytes(); - - // repeated .hbase.pb.TableServerTimestamp tst_map = 9; - /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ java.util.List<org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp> getTstMapList(); /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp getTstMap(int index); /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ int getTstMapCount(); /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestampOrBuilder> getTstMapOrBuilderList(); /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestampOrBuilder getTstMapOrBuilder( int index); - // repeated .hbase.pb.BackupImage dependent_backup_image = 10; + // repeated .hbase.pb.BackupImage dependent_backup_image = 8; /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ java.util.List<org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage> getDependentBackupImageList(); /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage getDependentBackupImage(int index); /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ int getDependentBackupImageCount(); /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImageOrBuilder> getDependentBackupImageOrBuilderList(); /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImageOrBuilder getDependentBackupImageOrBuilder( int index); - - // required bool compacted = 11; - /** - * <code>required bool compacted = 11;</code> - */ - boolean hasCompacted(); - /** - * <code>required bool compacted = 11;</code> - */ - boolean getCompacted(); } /** * Protobuf type {@code hbase.pb.BackupManifest} @@ -4582,37 +4552,22 @@ public final class BackupProtos { completeTs_ = input.readUInt64(); break; } - case 56: { - bitField0_ |= 0x00000020; - totalBytes_ = input.readInt64(); - break; - } - case 64: { - bitField0_ |= 0x00000040; - logBytes_ = input.readInt64(); - break; - } - case 74: { - if (!((mutable_bitField0_ & 0x00000100) == 0x00000100)) { + case 58: { + if (!((mutable_bitField0_ & 0x00000040) == 0x00000040)) { tstMap_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp>(); - mutable_bitField0_ |= 0x00000100; + mutable_bitField0_ |= 0x00000040; } tstMap_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp.PARSER, extensionRegistry)); break; } - case 82: { - if (!((mutable_bitField0_ & 0x00000200) == 0x00000200)) { + case 66: { + if (!((mutable_bitField0_ & 0x00000080) == 0x00000080)) { dependentBackupImage_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage>(); - mutable_bitField0_ |= 0x00000200; + mutable_bitField0_ |= 0x00000080; } dependentBackupImage_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage.PARSER, extensionRegistry)); break; } - case 88: { - bitField0_ |= 0x00000080; - compacted_ = input.readBool(); - break; - } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { @@ -4624,10 +4579,10 @@ public final class BackupProtos { if (((mutable_bitField0_ & 0x00000008) == 0x00000008)) { tableList_ = java.util.Collections.unmodifiableList(tableList_); } - if (((mutable_bitField0_ & 0x00000100) == 0x00000100)) { + if (((mutable_bitField0_ & 0x00000040) == 0x00000040)) { tstMap_ = java.util.Collections.unmodifiableList(tstMap_); } - if (((mutable_bitField0_ & 0x00000200) == 0x00000200)) { + if (((mutable_bitField0_ & 0x00000080) == 0x00000080)) { dependentBackupImage_ = java.util.Collections.unmodifiableList(dependentBackupImage_); } this.unknownFields = unknownFields.build(); @@ -4832,126 +4787,78 @@ public final class BackupProtos { return completeTs_; } - // required int64 total_bytes = 7; - public static final int TOTAL_BYTES_FIELD_NUMBER = 7; - private long totalBytes_; - /** - * <code>required int64 total_bytes = 7;</code> - */ - public boolean hasTotalBytes() { - return ((bitField0_ & 0x00000020) == 0x00000020); - } - /** - * <code>required int64 total_bytes = 7;</code> - */ - public long getTotalBytes() { - return totalBytes_; - } - - // optional int64 log_bytes = 8; - public static final int LOG_BYTES_FIELD_NUMBER = 8; - private long logBytes_; - /** - * <code>optional int64 log_bytes = 8;</code> - */ - public boolean hasLogBytes() { - return ((bitField0_ & 0x00000040) == 0x00000040); - } - /** - * <code>optional int64 log_bytes = 8;</code> - */ - public long getLogBytes() { - return logBytes_; - } - - // repeated .hbase.pb.TableServerTimestamp tst_map = 9; - public static final int TST_MAP_FIELD_NUMBER = 9; + // repeated .hbase.pb.TableServerTimestamp tst_map = 7; + public static final int TST_MAP_FIELD_NUMBER = 7; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp> tstMap_; /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp> getTstMapList() { return tstMap_; } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestampOrBuilder> getTstMapOrBuilderList() { return tstMap_; } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public int getTstMapCount() { return tstMap_.size(); } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp getTstMap(int index) { return tstMap_.get(index); } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestampOrBuilder getTstMapOrBuilder( int index) { return tstMap_.get(index); } - // repeated .hbase.pb.BackupImage dependent_backup_image = 10; - public static final int DEPENDENT_BACKUP_IMAGE_FIELD_NUMBER = 10; + // repeated .hbase.pb.BackupImage dependent_backup_image = 8; + public static final int DEPENDENT_BACKUP_IMAGE_FIELD_NUMBER = 8; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage> dependentBackupImage_; /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage> getDependentBackupImageList() { return dependentBackupImage_; } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImageOrBuilder> getDependentBackupImageOrBuilderList() { return dependentBackupImage_; } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public int getDependentBackupImageCount() { return dependentBackupImage_.size(); } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage getDependentBackupImage(int index) { return dependentBackupImage_.get(index); } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImageOrBuilder getDependentBackupImageOrBuilder( int index) { return dependentBackupImage_.get(index); } - // required bool compacted = 11; - public static final int COMPACTED_FIELD_NUMBER = 11; - private boolean compacted_; - /** - * <code>required bool compacted = 11;</code> - */ - public boolean hasCompacted() { - return ((bitField0_ & 0x00000080) == 0x00000080); - } - /** - * <code>required bool compacted = 11;</code> - */ - public boolean getCompacted() { - return compacted_; - } - private void initFields() { version_ = ""; backupId_ = ""; @@ -4959,11 +4866,8 @@ public final class BackupProtos { tableList_ = java.util.Collections.emptyList(); startTs_ = 0L; completeTs_ = 0L; - totalBytes_ = 0L; - logBytes_ = 0L; tstMap_ = java.util.Collections.emptyList(); dependentBackupImage_ = java.util.Collections.emptyList(); - compacted_ = false; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -4990,14 +4894,6 @@ public final class BackupProtos { memoizedIsInitialized = 0; return false; } - if (!hasTotalBytes()) { - memoizedIsInitialized = 0; - return false; - } - if (!hasCompacted()) { - memoizedIsInitialized = 0; - return false; - } for (int i = 0; i < getTableListCount(); i++) { if (!getTableList(i).isInitialized()) { memoizedIsInitialized = 0; @@ -5041,20 +4937,11 @@ public final class BackupProtos { if (((bitField0_ & 0x00000010) == 0x00000010)) { output.writeUInt64(6, completeTs_); } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - output.writeInt64(7, totalBytes_); - } - if (((bitField0_ & 0x00000040) == 0x00000040)) { - output.writeInt64(8, logBytes_); - } for (int i = 0; i < tstMap_.size(); i++) { - output.writeMessage(9, tstMap_.get(i)); + output.writeMessage(7, tstMap_.get(i)); } for (int i = 0; i < dependentBackupImage_.size(); i++) { - output.writeMessage(10, dependentBackupImage_.get(i)); - } - if (((bitField0_ & 0x00000080) == 0x00000080)) { - output.writeBool(11, compacted_); + output.writeMessage(8, dependentBackupImage_.get(i)); } getUnknownFields().writeTo(output); } @@ -5089,25 +4976,13 @@ public final class BackupProtos { size += com.google.protobuf.CodedOutputStream .computeUInt64Size(6, completeTs_); } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - size += com.google.protobuf.CodedOutputStream - .computeInt64Size(7, totalBytes_); - } - if (((bitField0_ & 0x00000040) == 0x00000040)) { - size += com.google.protobuf.CodedOutputStream - .computeInt64Size(8, logBytes_); - } for (int i = 0; i < tstMap_.size(); i++) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(9, tstMap_.get(i)); + .computeMessageSize(7, tstMap_.get(i)); } for (int i = 0; i < dependentBackupImage_.size(); i++) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(10, dependentBackupImage_.get(i)); - } - if (((bitField0_ & 0x00000080) == 0x00000080)) { - size += com.google.protobuf.CodedOutputStream - .computeBoolSize(11, compacted_); + .computeMessageSize(8, dependentBackupImage_.get(i)); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; @@ -5159,25 +5034,10 @@ public final class BackupProtos { result = result && (getCompleteTs() == other.getCompleteTs()); } - result = result && (hasTotalBytes() == other.hasTotalBytes()); - if (hasTotalBytes()) { - result = result && (getTotalBytes() - == other.getTotalBytes()); - } - result = result && (hasLogBytes() == other.hasLogBytes()); - if (hasLogBytes()) { - result = result && (getLogBytes() - == other.getLogBytes()); - } result = result && getTstMapList() .equals(other.getTstMapList()); result = result && getDependentBackupImageList() .equals(other.getDependentBackupImageList()); - result = result && (hasCompacted() == other.hasCompacted()); - if (hasCompacted()) { - result = result && (getCompacted() - == other.getCompacted()); - } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; @@ -5215,14 +5075,6 @@ public final class BackupProtos { hash = (37 * hash) + COMPLETE_TS_FIELD_NUMBER; hash = (53 * hash) + hashLong(getCompleteTs()); } - if (hasTotalBytes()) { - hash = (37 * hash) + TOTAL_BYTES_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getTotalBytes()); - } - if (hasLogBytes()) { - hash = (37 * hash) + LOG_BYTES_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getLogBytes()); - } if (getTstMapCount() > 0) { hash = (37 * hash) + TST_MAP_FIELD_NUMBER; hash = (53 * hash) + getTstMapList().hashCode(); @@ -5231,10 +5083,6 @@ public final class BackupProtos { hash = (37 * hash) + DEPENDENT_BACKUP_IMAGE_FIELD_NUMBER; hash = (53 * hash) + getDependentBackupImageList().hashCode(); } - if (hasCompacted()) { - hash = (37 * hash) + COMPACTED_FIELD_NUMBER; - hash = (53 * hash) + hashBoolean(getCompacted()); - } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; @@ -5363,24 +5211,18 @@ public final class BackupProtos { bitField0_ = (bitField0_ & ~0x00000010); completeTs_ = 0L; bitField0_ = (bitField0_ & ~0x00000020); - totalBytes_ = 0L; - bitField0_ = (bitField0_ & ~0x00000040); - logBytes_ = 0L; - bitField0_ = (bitField0_ & ~0x00000080); if (tstMapBuilder_ == null) { tstMap_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000100); + bitField0_ = (bitField0_ & ~0x00000040); } else { tstMapBuilder_.clear(); } if (dependentBackupImageBuilder_ == null) { dependentBackupImage_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000200); + bitField0_ = (bitField0_ & ~0x00000080); } else { dependentBackupImageBuilder_.clear(); } - compacted_ = false; - bitField0_ = (bitField0_ & ~0x00000400); return this; } @@ -5438,36 +5280,24 @@ public final class BackupProtos { to_bitField0_ |= 0x00000010; } result.completeTs_ = completeTs_; - if (((from_bitField0_ & 0x00000040) == 0x00000040)) { - to_bitField0_ |= 0x00000020; - } - result.totalBytes_ = totalBytes_; - if (((from_bitField0_ & 0x00000080) == 0x00000080)) { - to_bitField0_ |= 0x00000040; - } - result.logBytes_ = logBytes_; if (tstMapBuilder_ == null) { - if (((bitField0_ & 0x00000100) == 0x00000100)) { + if (((bitField0_ & 0x00000040) == 0x00000040)) { tstMap_ = java.util.Collections.unmodifiableList(tstMap_); - bitField0_ = (bitField0_ & ~0x00000100); + bitField0_ = (bitField0_ & ~0x00000040); } result.tstMap_ = tstMap_; } else { result.tstMap_ = tstMapBuilder_.build(); } if (dependentBackupImageBuilder_ == null) { - if (((bitField0_ & 0x00000200) == 0x00000200)) { + if (((bitField0_ & 0x00000080) == 0x00000080)) { dependentBackupImage_ = java.util.Collections.unmodifiableList(dependentBackupImage_); - bitField0_ = (bitField0_ & ~0x00000200); + bitField0_ = (bitField0_ & ~0x00000080); } result.dependentBackupImage_ = dependentBackupImage_; } else { result.dependentBackupImage_ = dependentBackupImageBuilder_.build(); } - if (((from_bitField0_ & 0x00000400) == 0x00000400)) { - to_bitField0_ |= 0x00000080; - } - result.compacted_ = compacted_; result.bitField0_ = to_bitField0_; onBuilt(); return result; @@ -5529,17 +5359,11 @@ public final class BackupProtos { if (other.hasCompleteTs()) { setCompleteTs(other.getCompleteTs()); } - if (other.hasTotalBytes()) { - setTotalBytes(other.getTotalBytes()); - } - if (other.hasLogBytes()) { - setLogBytes(other.getLogBytes()); - } if (tstMapBuilder_ == null) { if (!other.tstMap_.isEmpty()) { if (tstMap_.isEmpty()) { tstMap_ = other.tstMap_; - bitField0_ = (bitField0_ & ~0x00000100); + bitField0_ = (bitField0_ & ~0x00000040); } else { ensureTstMapIsMutable(); tstMap_.addAll(other.tstMap_); @@ -5552,7 +5376,7 @@ public final class BackupProtos { tstMapBuilder_.dispose(); tstMapBuilder_ = null; tstMap_ = other.tstMap_; - bitField0_ = (bitField0_ & ~0x00000100); + bitField0_ = (bitField0_ & ~0x00000040); tstMapBuilder_ = com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getTstMapFieldBuilder() : null; @@ -5565,7 +5389,7 @@ public final class BackupProtos { if (!other.dependentBackupImage_.isEmpty()) { if (dependentBackupImage_.isEmpty()) { dependentBackupImage_ = other.dependentBackupImage_; - bitField0_ = (bitField0_ & ~0x00000200); + bitField0_ = (bitField0_ & ~0x00000080); } else { ensureDependentBackupImageIsMutable(); dependentBackupImage_.addAll(other.dependentBackupImage_); @@ -5578,7 +5402,7 @@ public final class BackupProtos { dependentBackupImageBuilder_.dispose(); dependentBackupImageBuilder_ = null; dependentBackupImage_ = other.dependentBackupImage_; - bitField0_ = (bitField0_ & ~0x00000200); + bitField0_ = (bitField0_ & ~0x00000080); dependentBackupImageBuilder_ = com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getDependentBackupImageFieldBuilder() : null; @@ -5587,9 +5411,6 @@ public final class BackupProtos { } } } - if (other.hasCompacted()) { - setCompacted(other.getCompacted()); - } this.mergeUnknownFields(other.getUnknownFields()); return this; } @@ -5615,14 +5436,6 @@ public final class BackupProtos { return false; } - if (!hasTotalBytes()) { - - return false; - } - if (!hasCompacted()) { - - return false; - } for (int i = 0; i < getTableListCount(); i++) { if (!getTableList(i).isInitialized()) { @@ -6153,79 +5966,13 @@ public final class BackupProtos { return this; } - // required int64 total_bytes = 7; - private long totalBytes_ ; - /** - * <code>required int64 total_bytes = 7;</code> - */ - public boolean hasTotalBytes() { - return ((bitField0_ & 0x00000040) == 0x00000040); - } - /** - * <code>required int64 total_bytes = 7;</code> - */ - public long getTotalBytes() { - return totalBytes_; - } - /** - * <code>required int64 total_bytes = 7;</code> - */ - public Builder setTotalBytes(long value) { - bitField0_ |= 0x00000040; - totalBytes_ = value; - onChanged(); - return this; - } - /** - * <code>required int64 total_bytes = 7;</code> - */ - public Builder clearTotalBytes() { - bitField0_ = (bitField0_ & ~0x00000040); - totalBytes_ = 0L; - onChanged(); - return this; - } - - // optional int64 log_bytes = 8; - private long logBytes_ ; - /** - * <code>optional int64 log_bytes = 8;</code> - */ - public boolean hasLogBytes() { - return ((bitField0_ & 0x00000080) == 0x00000080); - } - /** - * <code>optional int64 log_bytes = 8;</code> - */ - public long getLogBytes() { - return logBytes_; - } - /** - * <code>optional int64 log_bytes = 8;</code> - */ - public Builder setLogBytes(long value) { - bitField0_ |= 0x00000080; - logBytes_ = value; - onChanged(); - return this; - } - /** - * <code>optional int64 log_bytes = 8;</code> - */ - public Builder clearLogBytes() { - bitField0_ = (bitField0_ & ~0x00000080); - logBytes_ = 0L; - onChanged(); - return this; - } - - // repeated .hbase.pb.TableServerTimestamp tst_map = 9; + // repeated .hbase.pb.TableServerTimestamp tst_map = 7; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp> tstMap_ = java.util.Collections.emptyList(); private void ensureTstMapIsMutable() { - if (!((bitField0_ & 0x00000100) == 0x00000100)) { + if (!((bitField0_ & 0x00000040) == 0x00000040)) { tstMap_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp>(tstMap_); - bitField0_ |= 0x00000100; + bitField0_ |= 0x00000040; } } @@ -6233,7 +5980,7 @@ public final class BackupProtos { org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp.Builder, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestampOrBuilder> tstMapBuilder_; /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp> getTstMapList() { if (tstMapBuilder_ == null) { @@ -6243,7 +5990,7 @@ public final class BackupProtos { } } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public int getTstMapCount() { if (tstMapBuilder_ == null) { @@ -6253,7 +6000,7 @@ public final class BackupProtos { } } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp getTstMap(int index) { if (tstMapBuilder_ == null) { @@ -6263,7 +6010,7 @@ public final class BackupProtos { } } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public Builder setTstMap( int index, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp value) { @@ -6280,7 +6027,7 @@ public final class BackupProtos { return this; } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public Builder setTstMap( int index, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp.Builder builderForValue) { @@ -6294,7 +6041,7 @@ public final class BackupProtos { return this; } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public Builder addTstMap(org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp value) { if (tstMapBuilder_ == null) { @@ -6310,7 +6057,7 @@ public final class BackupProtos { return this; } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public Builder addTstMap( int index, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp value) { @@ -6327,7 +6074,7 @@ public final class BackupProtos { return this; } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public Builder addTstMap( org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp.Builder builderForValue) { @@ -6341,7 +6088,7 @@ public final class BackupProtos { return this; } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public Builder addTstMap( int index, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp.Builder builderForValue) { @@ -6355,7 +6102,7 @@ public final class BackupProtos { return this; } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public Builder addAllTstMap( java.lang.Iterable<? extends org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp> values) { @@ -6369,12 +6116,12 @@ public final class BackupProtos { return this; } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public Builder clearTstMap() { if (tstMapBuilder_ == null) { tstMap_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000100); + bitField0_ = (bitField0_ & ~0x00000040); onChanged(); } else { tstMapBuilder_.clear(); @@ -6382,7 +6129,7 @@ public final class BackupProtos { return this; } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public Builder removeTstMap(int index) { if (tstMapBuilder_ == null) { @@ -6395,14 +6142,14 @@ public final class BackupProtos { return this; } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp.Builder getTstMapBuilder( int index) { return getTstMapFieldBuilder().getBuilder(index); } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestampOrBuilder getTstMapOrBuilder( int index) { @@ -6412,7 +6159,7 @@ public final class BackupProtos { } } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestampOrBuilder> getTstMapOrBuilderList() { @@ -6423,14 +6170,14 @@ public final class BackupProtos { } } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp.Builder addTstMapBuilder() { return getTstMapFieldBuilder().addBuilder( org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp.getDefaultInstance()); } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp.Builder addTstMapBuilder( int index) { @@ -6438,7 +6185,7 @@ public final class BackupProtos { index, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp.getDefaultInstance()); } /** - * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 9;</code> + * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 7;</code> */ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp.Builder> getTstMapBuilderList() { @@ -6451,7 +6198,7 @@ public final class BackupProtos { tstMapBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestamp.Builder, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.TableServerTimestampOrBuilder>( tstMap_, - ((bitField0_ & 0x00000100) == 0x00000100), + ((bitField0_ & 0x00000040) == 0x00000040), getParentForChildren(), isClean()); tstMap_ = null; @@ -6459,13 +6206,13 @@ public final class BackupProtos { return tstMapBuilder_; } - // repeated .hbase.pb.BackupImage dependent_backup_image = 10; + // repeated .hbase.pb.BackupImage dependent_backup_image = 8; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage> dependentBackupImage_ = java.util.Collections.emptyList(); private void ensureDependentBackupImageIsMutable() { - if (!((bitField0_ & 0x00000200) == 0x00000200)) { + if (!((bitField0_ & 0x00000080) == 0x00000080)) { dependentBackupImage_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage>(dependentBackupImage_); - bitField0_ |= 0x00000200; + bitField0_ |= 0x00000080; } } @@ -6473,7 +6220,7 @@ public final class BackupProtos { org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage.Builder, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImageOrBuilder> dependentBackupImageBuilder_; /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage> getDependentBackupImageList() { if (dependentBackupImageBuilder_ == null) { @@ -6483,7 +6230,7 @@ public final class BackupProtos { } } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public int getDependentBackupImageCount() { if (dependentBackupImageBuilder_ == null) { @@ -6493,7 +6240,7 @@ public final class BackupProtos { } } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage getDependentBackupImage(int index) { if (dependentBackupImageBuilder_ == null) { @@ -6503,7 +6250,7 @@ public final class BackupProtos { } } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public Builder setDependentBackupImage( int index, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage value) { @@ -6520,7 +6267,7 @@ public final class BackupProtos { return this; } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public Builder setDependentBackupImage( int index, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage.Builder builderForValue) { @@ -6534,7 +6281,7 @@ public final class BackupProtos { return this; } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public Builder addDependentBackupImage(org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage value) { if (dependentBackupImageBuilder_ == null) { @@ -6550,7 +6297,7 @@ public final class BackupProtos { return this; } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public Builder addDependentBackupImage( int index, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage value) { @@ -6567,7 +6314,7 @@ public final class BackupProtos { return this; } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public Builder addDependentBackupImage( org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage.Builder builderForValue) { @@ -6581,7 +6328,7 @@ public final class BackupProtos { return this; } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public Builder addDependentBackupImage( int index, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage.Builder builderForValue) { @@ -6595,7 +6342,7 @@ public final class BackupProtos { return this; } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public Builder addAllDependentBackupImage( java.lang.Iterable<? extends org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage> values) { @@ -6609,12 +6356,12 @@ public final class BackupProtos { return this; } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public Builder clearDependentBackupImage() { if (dependentBackupImageBuilder_ == null) { dependentBackupImage_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000200); + bitField0_ = (bitField0_ & ~0x00000080); onChanged(); } else { dependentBackupImageBuilder_.clear(); @@ -6622,7 +6369,7 @@ public final class BackupProtos { return this; } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public Builder removeDependentBackupImage(int index) { if (dependentBackupImageBuilder_ == null) { @@ -6635,14 +6382,14 @@ public final class BackupProtos { return this; } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage.Builder getDependentBackupImageBuilder( int index) { return getDependentBackupImageFieldBuilder().getBuilder(index); } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImageOrBuilder getDependentBackupImageOrBuilder( int index) { @@ -6652,7 +6399,7 @@ public final class BackupProtos { } } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImageOrBuilder> getDependentBackupImageOrBuilderList() { @@ -6663,14 +6410,14 @@ public final class BackupProtos { } } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage.Builder addDependentBackupImageBuilder() { return getDependentBackupImageFieldBuilder().addBuilder( org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage.getDefaultInstance()); } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage.Builder addDependentBackupImageBuilder( int index) { @@ -6678,7 +6425,7 @@ public final class BackupProtos { index, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage.getDefaultInstance()); } /** - * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 10;</code> + * <code>repeated .hbase.pb.BackupImage dependent_backup_image = 8;</code> */ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage.Builder> getDependentBackupImageBuilderList() { @@ -6691,7 +6438,7 @@ public final class BackupProtos { dependentBackupImageBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImage.Builder, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupImageOrBuilder>( dependentBackupImage_, - ((bitField0_ & 0x00000200) == 0x00000200), + ((bitField0_ & 0x00000080) == 0x00000080), getParentForChildren(), isClean()); dependentBackupImage_ = null; @@ -6699,39 +6446,6 @@ public final class BackupProtos { return dependentBackupImageBuilder_; } - // required bool compacted = 11; - private boolean compacted_ ; - /** - * <code>required bool compacted = 11;</code> - */ - public boolean hasCompacted() { - return ((bitField0_ & 0x00000400) == 0x00000400); - } - /** - * <code>required bool compacted = 11;</code> - */ - public boolean getCompacted() { - return compacted_; - } - /** - * <code>required bool compacted = 11;</code> - */ - public Builder setCompacted(boolean value) { - bitField0_ |= 0x00000400; - compacted_ = value; - onChanged(); - return this; - } - /** - * <code>required bool compacted = 11;</code> - */ - public Builder clearCompacted() { - bitField0_ = (bitField0_ & ~0x00000400); - compacted_ = false; - onChanged(); - return this; - } - // @@protoc_insertion_point(builder_scope:hbase.pb.BackupManifest) } @@ -7642,7 +7356,7 @@ public final class BackupProtos { // @@protoc_insertion_point(class_scope:hbase.pb.TableBackupStatus) } - public interface BackupContextOrBuilder + public interface BackupInfoOrBuilder extends com.google.protobuf.MessageOrBuilder { // required string backup_id = 1; @@ -7685,25 +7399,25 @@ public final class BackupProtos { com.google.protobuf.ByteString getTargetRootDirBytes(); - // optional .hbase.pb.BackupContext.BackupState state = 4; + // optional .hbase.pb.BackupInfo.BackupState state = 4; /** - * <code>optional .hbase.pb.BackupContext.BackupState state = 4;</code> + * <code>optional .hbase.pb.BackupInfo.BackupState state = 4;</code> */ boolean hasState(); /** - * <code>optional .hbase.pb.BackupContext.BackupState state = 4;</code> + * <code>optional .hbase.pb.BackupInfo.BackupState state = 4;</code> */ - org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupState getState(); + org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupState getState(); - // optional .hbase.pb.BackupContext.BackupPhase phase = 5; + // optional .hbase.pb.BackupInfo.BackupPhase phase = 5; /** - * <code>optional .hbase.pb.BackupContext.BackupPhase phase = 5;</code> + * <code>optional .hbase.pb.BackupInfo.BackupPhase phase = 5;</code> */ boolean hasPhase(); /** - * <code>optional .hbase.pb.BackupContext.BackupPhase phase = 5;</code> + * <code>optional .hbase.pb.BackupInfo.BackupPhase phase = 5;</code> */ - org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupPhase getPhase(); + org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupPhase getPhase(); // optional string failed_message = 6; /** @@ -7765,60 +7479,70 @@ public final class BackupProtos { */ long getEndTs(); - // optional int64 total_bytes_copied = 10; + // optional uint32 progress = 10; /** - * <code>optional int64 total_bytes_copied = 10;</code> + * <code>optional uint32 progress = 10;</code> */ - boolean hasTotalBytesCopied(); + boolean hasProgress(); /** - * <code>optional int64 total_bytes_copied = 10;</code> + * <code>optional uint32 progress = 10;</code> */ - long getTotalBytesCopied(); + int getProgress(); - // optional string hlog_target_dir = 11; + // optional string job_id = 11; /** - * <code>optional string hlog_target_dir = 11;</code> + * <code>optional string job_id = 11;</code> */ - boolean hasHlogTargetDir(); + boolean hasJobId(); /** - * <code>optional string hlog_target_dir = 11;</code> + * <code>optional string job_id = 11;</code> */ - java.lang.String getHlogTargetDir(); + java.lang.String getJobId(); /** - * <code>optional string hlog_target_dir = 11;</code> + * <code>optional string job_id = 11;</code> */ com.google.protobuf.ByteString - getHlogTargetDirBytes(); + getJobIdBytes(); - // optional uint32 progress = 12; + // required uint32 workers_number = 12; /** - * <code>optional uint32 progress = 12;</code> + * <code>required uint32 workers_number = 12;</code> */ - boolean hasProgress(); + boolean hasWorkersNumber(); /** - * <code>optional uint32 progress = 12;</code> + * <code>required uint32 workers_number = 12;</code> */ - int getProgress(); + int getWorkersNumber(); + + // required uint64 bandwidth = 13; + /** + * <code>required uint64 bandwidth = 13;</code> + */ + boolean hasBandwidth(); + /** + * <code>required uint64 bandwidth = 13;</code> + */ + long getBandwidth(); } /** - * Protobuf type {@code hbase.pb.BackupContext} + * Protobuf type {@code hbase.pb.BackupInfo} */ - public static final class BackupContext extends + public static final class BackupInfo extends com.google.protobuf.GeneratedMessage - implements BackupContextOrBuilder { - // Use BackupContext.newBuilder() to construct. - private BackupContext(com.google.protobuf.GeneratedMessage.Builder<?> builder) { + implements BackupInfoOrBuilder { + // Use BackupInfo.newBuilder() to construct. + private BackupInfo(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private BackupContext(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private BackupInfo(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - private static final BackupContext defaultInstance; - public static BackupContext getDefaultInstance() { + private static final BackupInfo defaultInstance; + public static BackupInfo getDefaultInstance() { return defaultInstance; } - public BackupContext getDefaultInstanceForType() { + public BackupInfo getDefaultInstanceForType() { return defaultInstance; } @@ -7828,7 +7552,7 @@ public final class BackupProtos { getUnknownFields() { return this.unknownFields; } - private BackupContext( + private BackupInfo( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { @@ -7874,7 +7598,7 @@ public final class BackupProtos { } case 32: { int rawValue = input.readEnum(); - org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupState value = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupState.valueOf(rawValue); + org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupState value = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupState.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(4, rawValue); } else { @@ -7885,7 +7609,7 @@ public final class BackupProtos { } case 40: { int rawValue = input.readEnum(); - org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupPhase value = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupPhase.valueOf(rawValue); + org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupPhase value = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupPhase.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(5, rawValue); } else { @@ -7919,17 +7643,22 @@ public final class BackupProtos { } case 80: { bitField0_ |= 0x00000100; - totalBytesCopied_ = input.readInt64(); + progress_ = input.readUInt32(); break; } case 90: { bitField0_ |= 0x00000200; - hlogTargetDir_ = input.readBytes(); + jobId_ = input.readBytes(); break; } case 96: { bitField0_ |= 0x00000400; - progress_ = input.readUInt32(); + workersNumber_ = input.readUInt32(); + break; + } + case 104: { + bitField0_ |= 0x00000800; + bandwidth_ = input.readUInt64(); break; } } @@ -7949,33 +7678,33 @@ public final class BackupProtos { } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.BackupProtos.internal_static_hbase_pb_BackupContext_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.BackupProtos.internal_static_hbase_pb_BackupInfo_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.BackupProtos.internal_static_hbase_pb_BackupContext_fieldAccessorTable + return org.apache.hadoop.hbase.protobuf.generated.BackupProtos.internal_static_hbase_pb_BackupInfo_fieldAccessorTable .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.class, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.Builder.class); + org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.class, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.Builder.class); } - public static com.google.protobuf.Parser<BackupContext> PARSER = - new com.google.protobuf.AbstractParser<BackupContext>() { - public BackupContext parsePartialFrom( + public static com.google.protobuf.Parser<BackupInfo> PARSER = + new com.google.protobuf.AbstractParser<BackupInfo>() { + public BackupInfo parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - return new BackupContext(input, extensionRegistry); + return new BackupInfo(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser<BackupContext> getParserForType() { + public com.google.protobuf.Parser<BackupInfo> getParserForType() { return PARSER; } /** - * Protobuf enum {@code hbase.pb.BackupContext.BackupState} + * Protobuf enum {@code hbase.pb.BackupInfo.BackupState} */ public enum BackupState implements com.google.protobuf.ProtocolMessageEnum { @@ -8058,7 +7787,7 @@ public final class BackupProtos { } public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.getDescriptor().getEnumTypes().get(0); + return org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.getDescriptor().getEnumTypes().get(0); } private static final BackupState[] VALUES = values(); @@ -8080,11 +7809,11 @@ public final class BackupProtos { this.value = value; } - // @@protoc_insertion_point(enum_scope:hbase.pb.BackupContext.BackupState) + // @@protoc_insertion_point(enum_scope:hbase.pb.BackupInfo.BackupState) } /** - * Protobuf enum {@code hbase.pb.BackupContext.BackupPhase} + * Protobuf enum {@code hbase.pb.BackupInfo.BackupPhase} */ public enum BackupPhase implements com.google.protobuf.ProtocolMessageEnum { @@ -8176,7 +7905,7 @@ public final class BackupProtos { } public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.getDescriptor().getEnumTypes().get(1); + return org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.getDescriptor().getEnumTypes().get(1); } private static final BackupPhase[] VALUES = values(); @@ -8198,7 +7927,7 @@ public final class BackupProtos { this.value = value; } - // @@protoc_insertion_point(enum_scope:hbase.pb.BackupContext.BackupPhase) + // @@protoc_insertion_point(enum_scope:hbase.pb.BackupInfo.BackupPhase) } private int bitField0_; @@ -8304,35 +8033,35 @@ public final class BackupProtos { } } - // optional .hbase.pb.BackupContext.BackupState state = 4; + // optional .hbase.pb.BackupInfo.BackupState state = 4; public static final int STATE_FIELD_NUMBER = 4; - private org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupState state_; + private org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupState state_; /** - * <code>optional .hbase.pb.BackupContext.BackupState state = 4;</code> + * <code>optional .hbase.pb.BackupInfo.BackupState state = 4;</code> */ public boolean hasState() { return ((bitField0_ & 0x00000008) == 0x00000008); } /** - * <code>optional .hbase.pb.BackupContext.BackupState state = 4;</code> + * <code>optional .hbase.pb.BackupInfo.BackupState state = 4;</code> */ - public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupState getState() { + public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupState getState() { return state_; } - // optional .hbase.pb.BackupContext.BackupPhase phase = 5; + // optional .hbase.pb.BackupInfo.BackupPhase phase = 5; public static final int PHASE_FIELD_NUMBER = 5; - private org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupPhase phase_; + private org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupPhase phase_; /** - * <code>optional .hbase.pb.BackupContext.BackupPhase phase = 5;</code> + * <code>optional .hbase.pb.BackupInfo.BackupPhase phase = 5;</code> */ public boolean hasPhase() { return ((bitField0_ & 0x00000010) == 0x00000010); } /** - * <code>optional .hbase.pb.BackupContext.BackupPhase phase = 5;</code> + * <code>optional .hbase.pb.BackupInfo.BackupPhase phase = 5;</code> */ - public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupPhase getPhase() { + public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupPhase getPhase() { return phase_; } @@ -8447,36 +8176,36 @@ public final class BackupProtos { return endTs_; } - // optional int64 total_bytes_copied = 10; - public static final int TOTAL_BYTES_COPIED_FIELD_NUMBER = 10; - private long totalBytesCopied_; + // optional uint32 progress = 10; + public static final int PROGRESS_FIELD_NUMBER = 10; + private int progress_; /** - * <code>optional int64 total_bytes_copied = 10;</code> + * <code>optional uint32 progress = 10;</code> */ - public boolean hasTotalBytesCopied() { + public boolean hasProgress() { return ((bitField0_ & 0x00000100) == 0x00000100); } /** - * <code>optional int64 total_bytes_copied = 10;</code> + * <code>optional uint32 progress = 10;</code> */ - public long getTotalBytesCopied() { - return totalBytesCopied_; + public int getProgress() { + return progress_; } - // optional string hlog_target_dir = 11; - public static final int HLOG_TARGET_DIR_FIELD_NUMBER = 11; - private java.lang.Object hlogTargetDir_; + // optional string job_id = 11; + public static final int JOB_ID_FIELD_NUMBER = 11; + private java.lang.Object jobId_; /** - * <code>optional string hlog_target_dir = 11;</code> + * <code>optional string job_id = 11;</code> */ - public boolean hasHlogTargetDir() { + public boolean hasJobId() { return ((bitField0_ & 0x00000200) == 0x00000200); } /** - * <code>optional string hlog_target_dir = 11;</code> + * <code>optional string job_id = 11;</code> */ - public java.lang.String getHlogTargetDir() { - java.lang.Object ref = hlogTargetDir_; + public java.lang.String getJobId() { + java.lang.Object ref = jobId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { @@ -8484,57 +8213,74 @@ public final class BackupProtos { (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { - hlogTargetDir_ = s; + jobId_ = s; } return s; } } /** - * <code>optional string hlog_target_dir = 11;</code> + * <code>optional string job_id = 11;</code> */ public com.google.protobuf.ByteString - getHlogTargetDirBytes() { - java.lang.Object ref = hlogTargetDir_; + getJobIdBytes() { + java.lang.Object ref = jobId_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); - hlogTargetDir_ = b; + jobId_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } - // optional uint32 progress = 12; - public static final int PROGRESS_FIELD_NUMBER = 12; - private int progress_; + // required uint32 workers_number = 12; + public static final int WORKERS_NUMBER_FIELD_NUMBER = 12; + private int workersNumber_; /** - * <code>optional uint32 progress = 12;</code> + * <code>required uint32 workers_number = 12;</code> */ - public boolean hasProgress() { + public boolean hasWorkersNumber() { return ((bitField0_ & 0x00000400) == 0x00000400); } /** - * <code>optional uint32 progress = 12;</code> + * <code>required uint32 workers_number = 12;</code> */ - public int getProgress() { - return progress_; + public int getWorkersNumber() { + return workersNumber_; + } + + // required uint64 bandwidth = 13; + public static final int BANDWIDTH_FIELD_NUMBER = 13; + private long bandwidth_; + /** + * <code>required uint64 bandwidth = 13;</code> + */ + public boolean hasBandwidth() { + return ((bitField0_ & 0x00000800) == 0x00000800); + } + /** + * <code>required uint64 bandwidth = 13;</code> + */ + public long getBandwidth() { + return bandwidth_; } private void initFields() { backupId_ = ""; type_ = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupType.FULL; targetRootDir_ = ""; - state_ = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupState.WAITING; - phase_ = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupPhase.REQUEST; + state_ = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupState.WAITING; + phase_ = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupPhase.REQUEST; failedMessage_ = ""; tableBackupStatus_ = java.util.Collections.emptyList(); startTs_ = 0L; endTs_ = 0L; - totalBytesCopied_ = 0L; - hlogTargetDir_ = ""; progress_ = 0; + jobId_ = ""; + workersNumber_ = 0; + bandwidth_ = 0L; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -8553,6 +8299,14 @@ public final class BackupProtos { memoizedIsInitialized = 0; return false; } + if (!hasWorkersNumber()) { + memoizedIsInitialized = 0; + return false; + } + if (!hasBandwidth()) { + memoizedIsInitialized = 0; + return false; + } for (int i = 0; i < getTableBackupStatusCount(); i++) { if (!getTableBackupStatus(i).isInitialized()) { memoizedIsInitialized = 0; @@ -8594,13 +8348,16 @@ public final class BackupProtos { output.writeUInt64(9, endTs_); } if (((bitField0_ & 0x00000100) == 0x00000100)) { - output.writeInt64(10, totalBytesCopied_); + output.writeUInt32(10, progress_); } if (((bitField0_ & 0x00000200) == 0x00000200)) { - output.writeBytes(11, getHlogTargetDirBytes()); + output.writeBytes(11, getJobIdBytes()); } if (((bitField0_ & 0x00000400) == 0x00000400)) { - output.writeUInt32(12, progress_); + output.writeUInt32(12, workersNumber_); + } + if (((bitField0_ & 0x00000800) == 0x00000800)) { + output.writeUInt64(13, bandwidth_); } getUnknownFields().writeTo(output); } @@ -8649,15 +8406,19 @@ public final class BackupProtos { } if (((bitField0_ & 0x00000100) == 0x00000100)) { size += com.google.protobuf.CodedOutputStream - .computeInt64Size(10, totalBytesCopied_); + .computeUInt32Size(10, progress_); } if (((bitField0_ & 0x00000200) == 0x00000200)) { size += com.google.protobuf.CodedOutputStream - .computeBytesSize(11, getHlogTargetDirBytes()); + .computeBytesSize(11, getJobIdBytes()); } if (((bitField0_ & 0x00000400) == 0x00000400)) { size += com.google.protobuf.CodedOutputStream - .computeUInt32Size(12, progress_); + .computeUInt32Size(12, workersNumber_); + } + if (((bitField0_ & 0x00000800) == 0x00000800)) { + size += com.google.protobuf.CodedOutputStream + .computeUInt64Size(13, bandwidth_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; @@ -8676,10 +8437,10 @@ public final class BackupProtos { if (obj == this) { return true; } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext)) { + if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo)) { return super.equals(obj); } - org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext other = (org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext) obj; + org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo other = (org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo) obj; boolean result = true; result = result && (hasBackupId() == other.hasBackupId()); @@ -8724,21 +8485,26 @@ public final class BackupProtos { result = result && (getEndTs() == other.getEndTs()); } - result = result && (hasTotalBytesCopied() == other.hasTotalBytesCopied()); - if (hasTotalBytesCopied()) { - result = result && (getTotalBytesCopied() - == other.getTotalBytesCopied()); - } - result = result && (hasHlogTargetDir() == other.hasHlogTargetDir()); - if (hasHlogTargetDir()) { - result = result && getHlogTargetDir() - .equals(other.getHlogTargetDir()); - } result = result && (hasProgress() == other.hasProgress()); if (hasProgress()) { result = result && (getProgress() == other.getProgress()); } + result = result && (hasJobId() == other.hasJobId()); + if (hasJobId()) { + result = result && getJobId() + .equals(other.getJobId()); + } + result = result && (hasWorkersNumber() == other.hasWorkersNumber()); + if (hasWorkersNumber()) { + result = result && (getWorkersNumber() + == other.getWorkersNumber()); + } + result = result && (hasBandwidth() == other.hasBandwidth()); + if (hasBandwidth()) { + result = result && (getBandwidth() + == other.getBandwidth()); + } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; @@ -8788,70 +8554,74 @@ public final class BackupProtos { hash = (37 * hash) + END_TS_FIELD_NUMBER; hash = (53 * hash) + hashLong(getEndTs()); } - if (hasTotalBytesCopied()) { - hash = (37 * hash) + TOTAL_BYTES_COPIED_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getTotalBytesCopied()); - } - if (hasHlogTargetDir()) { - hash = (37 * hash) + HLOG_TARGET_DIR_FIELD_NUMBER; - hash = (53 * hash) + getHlogTargetDir().hashCode(); - } if (hasProgress()) { hash = (37 * hash) + PROGRESS_FIELD_NUMBER; hash = (53 * hash) + getProgress(); } + if (hasJobId()) { + hash = (37 * hash) + JOB_ID_FIELD_NUMBER; + hash = (53 * hash) + getJobId().hashCode(); + } + if (hasWorkersNumber()) { + hash = (37 * hash) + WORKERS_NUMBER_FIELD_NUMBER; + hash = (53 * hash) + getWorkersNumber(); + } + if (hasBandwidth()) { + hash = (37 * hash) + BANDWIDTH_FIELD_NUMBER; + hash = (53 * hash) + hashLong(getBandwidth()); + } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } - public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } - public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } - public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext parseFrom(byte[] data) + public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } - public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } - public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext parseFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } - public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } - public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext parseDelimitedFrom(java.io.InputStream input) + public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } - public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext parseDelimitedFrom( + public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } - public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } - public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext parseFrom( + public static org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -8860,7 +8630,7 @@ public final class BackupProtos { public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext prototype) { + public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @@ -8872,24 +8642,24 @@ public final class BackupProtos { return builder; } /** - * Protobuf type {@code hbase.pb.BackupContext} + * Protobuf type {@code hbase.pb.BackupInfo} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> - implements org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContextOrBuilder { + implements org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.BackupProtos.internal_static_hbase_pb_BackupContext_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.BackupProtos.internal_static_hbase_pb_BackupInfo_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.BackupProtos.internal_static_hbase_pb_BackupContext_fieldAccessorTable + return org.apache.hadoop.hbase.protobuf.generated.BackupProtos.internal_static_hbase_pb_BackupInfo_fieldAccessorTable .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.class, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.Builder.class); + org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.class, org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.Builder.class); } - // Construct using org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.newBuilder() + // Construct using org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.newBuilder() private Builder() { maybeForceBuilderInitialization(); } @@ -8916,9 +8686,9 @@ public final class BackupProtos { bitField0_ = (bitField0_ & ~0x00000002); targetRootDir_ = ""; bitField0_ = (bitField0_ & ~0x00000004); - state_ = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupState.WAITING; + state_ = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupState.WAITING; bitField0_ = (bitField0_ & ~0x00000008); - phase_ = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupPhase.REQUEST; + phase_ = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupPhase.REQUEST; bitField0_ = (bitField0_ & ~0x00000010); failedMessage_ = ""; bitField0_ = (bitField0_ & ~0x00000020); @@ -8932,12 +8702,14 @@ public final class BackupProtos { bitField0_ = (bitField0_ & ~0x00000080); endTs_ = 0L; bitField0_ = (bitField0_ & ~0x00000100); - totalBytesCopied_ = 0L; + progress_ = 0; bitField0_ = (bitField0_ & ~0x00000200); - hlogTargetDir_ = ""; + jobId_ = ""; bitField0_ = (bitField0_ & ~0x00000400); - progress_ = 0; + workersNumber_ = 0; bitField0_ = (bitField0_ & ~0x00000800); + bandwidth_ = 0L; + bitField0_ = (bitField0_ & ~0x00001000); return this; } @@ -8947,23 +8719,23 @@ public final class BackupProtos { public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.BackupProtos.internal_static_hbase_pb_BackupContext_descriptor; + return org.apache.hadoop.hbase.protobuf.generated.BackupProtos.internal_static_hbase_pb_BackupInfo_descriptor; } - public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.getDefaultInstance(); + public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo getDefaultInstanceForType() { + return org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.getDefaultInstance(); } - public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext build() { - org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext result = buildPartial(); + public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo build() { + org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } - public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext result = new org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext(this); + public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo buildPartial() { + org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo result = new org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { @@ -9010,31 +8782,35 @@ public final class BackupProtos { if (((from_bitField0_ & 0x00000200) == 0x00000200)) { to_bitField0_ |= 0x00000100; } - result.totalBytesCopied_ = totalBytesCopied_; + result.progress_ = progress_; if (((from_bitField0_ & 0x00000400) == 0x00000400)) { to_bitField0_ |= 0x00000200; } - result.hlogTargetDir_ = hlogTargetDir_; + result.jobId_ = jobId_; if (((from_bitField0_ & 0x00000800) == 0x00000800)) { to_bitField0_ |= 0x00000400; } - result.progress_ = progress_; + result.workersNumber_ = workersNumber_; + if (((from_bitField0_ & 0x00001000) == 0x00001000)) { + to_bitField0_ |= 0x00000800; + } + result.bandwidth_ = bandwidth_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext)other); + if (other instanceof org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo) { + return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo)other); } else { super.mergeFrom(other); return this; } } - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.getDefaultInstance()) return this; + public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo other) { + if (other == org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.getDefaultInstance()) return this; if (other.hasBackupId()) { bitField0_ |= 0x00000001; backupId_ = other.backupId_; @@ -9091,16 +8867,19 @@ public final class BackupProtos { if (other.hasEndTs()) { setEndTs(other.getEndTs()); } - if (other.hasTotalBytesCopied()) { - setTotalBytesCopied(other.getTotalBytesCopied()); + if (other.hasProgress()) { + setProgress(other.getProgress()); } - if (other.hasHlogTargetDir()) { + if (other.hasJobId()) { bitField0_ |= 0x00000400; - hlogTargetDir_ = other.hlogTargetDir_; + jobId_ = other.jobId_; onChanged(); } - if (other.hasProgress()) { - setProgress(other.getProgress()); + if (other.hasWorkersNumber()) { + setWorkersNumber(other.getWorkersNumber()); + } + if (other.hasBandwidth()) { + setBandwidth(other.getBandwidth()); } this.mergeUnknownFields(other.getUnknownFields()); return this; @@ -9119,6 +8898,14 @@ public final class BackupProtos { return false; } + if (!hasWorkersNumber()) { + + return false; + } + if (!hasBandwidth()) { + + return false; + } for (int i = 0; i < getTableBackupStatusCount(); i++) { if (!getTableBackupStatus(i).isInitialized()) { @@ -9132,11 +8919,11 @@ public final class BackupProtos { com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext parsedMessage = null; + org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext) e.getUnfinishedMessage(); + parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { @@ -9331,24 +9118,24 @@ public final class BackupProtos { return this; } - // optional .hbase.pb.BackupContext.BackupState state = 4; - private org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupState state_ = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupState.WAITING; + // optional .hbase.pb.BackupInfo.BackupState state = 4; + private org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupState state_ = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupState.WAITING; /** - * <code>optional .hbase.pb.BackupContext.BackupState state = 4;</code> + * <code>optional .hbase.pb.BackupInfo.BackupState state = 4;</code> */ public boolean hasState() { return ((bitField0_ & 0x00000008) == 0x00000008); } /** - * <code>optional .hbase.pb.BackupContext.BackupState state = 4;</code> + * <code>optional .hbase.pb.BackupInfo.BackupState state = 4;</code> */ - public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupState getState() { + public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupState getState() { return state_; } /** - * <code>optional .hbase.pb.BackupContext.BackupState state = 4;</code> + * <code>optional .hbase.pb.BackupInfo.BackupState state = 4;</code> */ - public Builder setState(org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupState value) { + public Builder setState(org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupState value) { if (value == null) { throw new NullPointerException(); } @@ -9358,33 +9145,33 @@ public final class BackupProtos { return this; } /** - * <code>optional .hbase.pb.BackupContext.BackupState state = 4;</code> + * <code>optional .hbase.pb.BackupInfo.BackupState state = 4;</code> */ public Builder clearState() { bitField0_ = (bitField0_ & ~0x00000008); - state_ = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupState.WAITING; + state_ = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupState.WAITING; onChanged(); return this; } - // optional .hbase.pb.BackupContext.BackupPhase phase = 5; - private org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupPhase phase_ = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupPhase.REQUEST; + // optional .hbase.pb.BackupInfo.BackupPhase phase = 5; + private org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupPhase phase_ = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupPhase.REQUEST; /** - * <code>optional .hbase.pb.BackupContext.BackupPhase phase = 5;</code> + * <code>optional .hbase.pb.BackupInfo.BackupPhase phase = 5;</code> */ public boolean hasPhase() { return ((bitField0_ & 0x00000010) == 0x00000010); } /** - * <code>optional .hbase.pb.BackupContext.BackupPhase phase = 5;</code> + * <code>optional .hbase.pb.BackupInfo.BackupPhase phase = 5;</code> */ - public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupPhase getPhase() { + public org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupPhase getPhase() { return phase_; } /** - * <code>optional .hbase.pb.BackupContext.BackupPhase phase = 5;</code> + * <code>optional .hbase.pb.BackupInfo.BackupPhase phase = 5;</code> */ - public Builder setPhase(org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupPhase value) { + public Builder setPhase(org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupPhase value) { if (value == null) { throw new NullPointerException(); } @@ -9394,11 +9181,11 @@ public final class BackupProtos { return this; } /** - * <code>optional .hbase.pb.BackupContext.BackupPhase phase = 5;</code> + * <code>optional .hbase.pb.BackupInfo.BackupPhase phase = 5;</code> */ public Builder clearPhase() { bitField0_ = (bitField0_ & ~0x00000010); - phase_ = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupContext.BackupPhase.REQUEST; + phase_ = org.apache.hadoop.hbase.protobuf.generated.BackupProtos.BackupInfo.BackupPhase.REQUEST; onChanged(); return this; } @@ -9783,173 +9570,206 @@ public final class BackupProtos { return this; } - // optional int64 total_bytes_copied = 10; - private long totalBytesCopied_ ; + // optional uint32 progress = 10; + private int progress_ ; /** - * <code>optional int64 total_bytes_copied = 10;</code> + * <code>optional uint32 progress = 10;</code> */ - public boolean hasTotalB
<TRUNCATED>