// Generated by the protocol buffer compiler. DO NOT EDIT!
// source: StorageClusterStatusMessage.proto
package org.apache.hadoop.hbase.rest.protobuf.generated;
public final class StorageClusterStatusMessage {
private StorageClusterStatusMessage() {}
public static void registerAllExtensions(
com.google.protobuf.ExtensionRegistry registry) {
}
public static final class StorageClusterStatus extends
com.google.protobuf.GeneratedMessage {
// Use StorageClusterStatus.newBuilder() to construct.
private StorageClusterStatus() {
initFields();
}
private StorageClusterStatus(boolean noInit) {}
private static final StorageClusterStatus defaultInstance;
public static StorageClusterStatus getDefaultInstance() {
return defaultInstance;
}
public StorageClusterStatus getDefaultInstanceForType() {
return defaultInstance;
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_fieldAccessorTable;
}
public static final class Region extends
com.google.protobuf.GeneratedMessage {
// Use Region.newBuilder() to construct.
private Region() {
initFields();
}
private Region(boolean noInit) {}
private static final Region defaultInstance;
public static Region getDefaultInstance() {
return defaultInstance;
}
public Region getDefaultInstanceForType() {
return defaultInstance;
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_fieldAccessorTable;
}
// required bytes name = 1;
public static final int NAME_FIELD_NUMBER = 1;
private boolean hasName;
private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY;
public boolean hasName() { return hasName; }
public com.google.protobuf.ByteString getName() { return name_; }
// optional int32 stores = 2;
public static final int STORES_FIELD_NUMBER = 2;
private boolean hasStores;
private int stores_ = 0;
public boolean hasStores() { return hasStores; }
public int getStores() { return stores_; }
// optional int32 storefiles = 3;
public static final int STOREFILES_FIELD_NUMBER = 3;
private boolean hasStorefiles;
private int storefiles_ = 0;
public boolean hasStorefiles() { return hasStorefiles; }
public int getStorefiles() { return storefiles_; }
// optional int32 storefileSizeMB = 4;
public static final int STOREFILESIZEMB_FIELD_NUMBER = 4;
private boolean hasStorefileSizeMB;
private int storefileSizeMB_ = 0;
public boolean hasStorefileSizeMB() { return hasStorefileSizeMB; }
public int getStorefileSizeMB() { return storefileSizeMB_; }
// optional int32 memstoreSizeMB = 5;
public static final int MEMSTORESIZEMB_FIELD_NUMBER = 5;
private boolean hasMemstoreSizeMB;
private int memstoreSizeMB_ = 0;
public boolean hasMemstoreSizeMB() { return hasMemstoreSizeMB; }
public int getMemstoreSizeMB() { return memstoreSizeMB_; }
// optional int32 storefileIndexSizeMB = 6;
public static final int STOREFILEINDEXSIZEMB_FIELD_NUMBER = 6;
private boolean hasStorefileIndexSizeMB;
private int storefileIndexSizeMB_ = 0;
public boolean hasStorefileIndexSizeMB() { return hasStorefileIndexSizeMB; }
public int getStorefileIndexSizeMB() { return storefileIndexSizeMB_; }
private void initFields() {
}
public final boolean isInitialized() {
if (!hasName) return false;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
if (hasName()) {
output.writeBytes(1, getName());
}
if (hasStores()) {
output.writeInt32(2, getStores());
}
if (hasStorefiles()) {
output.writeInt32(3, getStorefiles());
}
if (hasStorefileSizeMB()) {
output.writeInt32(4, getStorefileSizeMB());
}
if (hasMemstoreSizeMB()) {
output.writeInt32(5, getMemstoreSizeMB());
}
if (hasStorefileIndexSizeMB()) {
output.writeInt32(6, getStorefileIndexSizeMB());
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
if (hasName()) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(1, getName());
}
if (hasStores()) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(2, getStores());
}
if (hasStorefiles()) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(3, getStorefiles());
}
if (hasStorefileSizeMB()) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(4, getStorefileSizeMB());
}
if (hasMemstoreSizeMB()) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(5, getMemstoreSizeMB());
}
if (hasStorefileIndexSizeMB()) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(6, getStorefileIndexSizeMB());
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data).buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data, extensionRegistry)
.buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data).buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data, extensionRegistry)
.buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(java.io.InputStream input)
throws java.io.IOException {
return newBuilder().mergeFrom(input).buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return newBuilder().mergeFrom(input, extensionRegistry)
.buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
Builder builder = newBuilder();
if (builder.mergeDelimitedFrom(input)) {
return builder.buildParsed();
} else {
return null;
}
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
Builder builder = newBuilder();
if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
return builder.buildParsed();
} else {
return null;
}
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return newBuilder().mergeFrom(input).buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return newBuilder().mergeFrom(input, extensionRegistry)
.buildParsed();
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder<Builder> {
private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region result;
// Construct using org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.newBuilder()
private Builder() {}
private static Builder create() {
Builder builder = new Builder();
builder.result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region();
return builder;
}
protected org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region internalGetResult() {
return result;
}
public Builder clear() {
if (result == null) {
throw new IllegalStateException(
"Cannot call clear() after build().");
}
result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region();
return this;
}
public Builder clone() {
return create().mergeFrom(result);
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.getDescriptor();
}
public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region getDefaultInstanceForType() {
return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.getDefaultInstance();
}
public boolean isInitialized() {
return result.isInitialized();
}
public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region build() {
if (result != null && !isInitialized()) {
throw newUninitializedMessageException(result);
}
return buildPartial();
}
private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region buildParsed()
throws com.google.protobuf.InvalidProtocolBufferException {
if (!isInitialized()) {
throw newUninitializedMessageException(
result).asInvalidProtocolBufferException();
}
return buildPartial();
}
public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region buildPartial() {
if (result == null) {
throw new IllegalStateException(
"build() has already been called on this Builder.");
}
org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region returnMe = result;
result = null;
return returnMe;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region) {
return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region other) {
if (other == org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.getDefaultInstance()) return this;
if (other.hasName()) {
setName(other.getName());
}
if (other.hasStores()) {
setStores(other.getStores());
}
if (other.hasStorefiles()) {
setStorefiles(other.getStorefiles());
}
if (other.hasStorefileSizeMB()) {
setStorefileSizeMB(other.getStorefileSizeMB());
}
if (other.hasMemstoreSizeMB()) {
setMemstoreSizeMB(other.getMemstoreSizeMB());
}
if (other.hasStorefileIndexSizeMB()) {
setStorefileIndexSizeMB(other.getStorefileIndexSizeMB());
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder(
this.getUnknownFields());
while (true) {
int tag = input.readTag();
switch (tag) {
case 0:
this.setUnknownFields(unknownFields.build());
return this;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
this.setUnknownFields(unknownFields.build());
return this;
}
break;
}
case 10: {
setName(input.readBytes());
break;
}
case 16: {
setStores(input.readInt32());
break;
}
case 24: {
setStorefiles(input.readInt32());
break;
}
case 32: {
setStorefileSizeMB(input.readInt32());
break;
}
case 40: {
setMemstoreSizeMB(input.readInt32());
break;
}
case 48: {
setStorefileIndexSizeMB(input.readInt32());
break;
}
}
}
}
// required bytes name = 1;
public boolean hasName() {
return result.hasName();
}
public com.google.protobuf.ByteString getName() {
return result.getName();
}
public Builder setName(com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
result.hasName = true;
result.name_ = value;
return this;
}
public Builder clearName() {
result.hasName = false;
result.name_ = getDefaultInstance().getName();
return this;
}
// optional int32 stores = 2;
public boolean hasStores() {
return result.hasStores();
}
public int getStores() {
return result.getStores();
}
public Builder setStores(int value) {
result.hasStores = true;
result.stores_ = value;
return this;
}
public Builder clearStores() {
result.hasStores = false;
result.stores_ = 0;
return this;
}
// optional int32 storefiles = 3;
public boolean hasStorefiles() {
return result.hasStorefiles();
}
public int getStorefiles() {
return result.getStorefiles();
}
public Builder setStorefiles(int value) {
result.hasStorefiles = true;
result.storefiles_ = value;
return this;
}
public Builder clearStorefiles() {
result.hasStorefiles = false;
result.storefiles_ = 0;
return this;
}
// optional int32 storefileSizeMB = 4;
public boolean hasStorefileSizeMB() {
return result.hasStorefileSizeMB();
}
public int getStorefileSizeMB() {
return result.getStorefileSizeMB();
}
public Builder setStorefileSizeMB(int value) {
result.hasStorefileSizeMB = true;
result.storefileSizeMB_ = value;
return this;
}
public Builder clearStorefileSizeMB() {
result.hasStorefileSizeMB = false;
result.storefileSizeMB_ = 0;
return this;
}
// optional int32 memstoreSizeMB = 5;
public boolean hasMemstoreSizeMB() {
return result.hasMemstoreSizeMB();
}
public int getMemstoreSizeMB() {
return result.getMemstoreSizeMB();
}
public Builder setMemstoreSizeMB(int value) {
result.hasMemstoreSizeMB = true;
result.memstoreSizeMB_ = value;
return this;
}
public Builder clearMemstoreSizeMB() {
result.hasMemstoreSizeMB = false;
result.memstoreSizeMB_ = 0;
return this;
}
// optional int32 storefileIndexSizeMB = 6;
public boolean hasStorefileIndexSizeMB() {
return result.hasStorefileIndexSizeMB();
}
public int getStorefileIndexSizeMB() {
return result.getStorefileIndexSizeMB();
}
public Builder setStorefileIndexSizeMB(int value) {
result.hasStorefileIndexSizeMB = true;
result.storefileIndexSizeMB_ = value;
return this;
}
public Builder clearStorefileIndexSizeMB() {
result.hasStorefileIndexSizeMB = false;
result.storefileIndexSizeMB_ = 0;
return this;
}
// @@protoc_insertion_point(builder_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Region)
}
static {
defaultInstance = new Region(true);
org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internalForceInit();
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Region)
}
public static final class Node extends
com.google.protobuf.GeneratedMessage {
// Use Node.newBuilder() to construct.
private Node() {
initFields();
}
private Node(boolean noInit) {}
private static final Node defaultInstance;
public static Node getDefaultInstance() {
return defaultInstance;
}
public Node getDefaultInstanceForType() {
return defaultInstance;
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_fieldAccessorTable;
}
// required string name = 1;
public static final int NAME_FIELD_NUMBER = 1;
private boolean hasName;
private java.lang.String name_ = "";
public boolean hasName() { return hasName; }
public java.lang.String getName() { return name_; }
// optional int64 startCode = 2;
public static final int STARTCODE_FIELD_NUMBER = 2;
private boolean hasStartCode;
private long startCode_ = 0L;
public boolean hasStartCode() { return hasStartCode; }
public long getStartCode() { return startCode_; }
// optional int32 requests = 3;
public static final int REQUESTS_FIELD_NUMBER = 3;
private boolean hasRequests;
private int requests_ = 0;
public boolean hasRequests() { return hasRequests; }
public int getRequests() { return requests_; }
// optional int32 heapSizeMB = 4;
public static final int HEAPSIZEMB_FIELD_NUMBER = 4;
private boolean hasHeapSizeMB;
private int heapSizeMB_ = 0;
public boolean hasHeapSizeMB() { return hasHeapSizeMB; }
public int getHeapSizeMB() { return heapSizeMB_; }
// optional int32 maxHeapSizeMB = 5;
public static final int MAXHEAPSIZEMB_FIELD_NUMBER = 5;
private boolean hasMaxHeapSizeMB;
private int maxHeapSizeMB_ = 0;
public boolean hasMaxHeapSizeMB() { return hasMaxHeapSizeMB; }
public int getMaxHeapSizeMB() { return maxHeapSizeMB_; }
// repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Region regions = 6;
public static final int REGIONS_FIELD_NUMBER = 6;
private java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> regions_ =
java.util.Collections.emptyList();
public java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> getRegionsList() {
return regions_;
}
public int getRegionsCount() { return regions_.size(); }
public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region getRegions(int index) {
return regions_.get(index);
}
private void initFields() {
}
public final boolean isInitialized() {
if (!hasName) return false;
for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region element : getRegionsList()) {
if (!element.isInitialized()) return false;
}
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
if (hasName()) {
output.writeString(1, getName());
}
if (hasStartCode()) {
output.writeInt64(2, getStartCode());
}
if (hasRequests()) {
output.writeInt32(3, getRequests());
}
if (hasHeapSizeMB()) {
output.writeInt32(4, getHeapSizeMB());
}
if (hasMaxHeapSizeMB()) {
output.writeInt32(5, getMaxHeapSizeMB());
}
for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region element : getRegionsList()) {
output.writeMessage(6, element);
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
if (hasName()) {
size += com.google.protobuf.CodedOutputStream
.computeStringSize(1, getName());
}
if (hasStartCode()) {
size += com.google.protobuf.CodedOutputStream
.computeInt64Size(2, getStartCode());
}
if (hasRequests()) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(3, getRequests());
}
if (hasHeapSizeMB()) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(4, getHeapSizeMB());
}
if (hasMaxHeapSizeMB()) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(5, getMaxHeapSizeMB());
}
for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region element : getRegionsList()) {
size += com.google.protobuf.CodedOutputStream
.computeMessageSize(6, element);
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data).buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data, extensionRegistry)
.buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data).buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data, extensionRegistry)
.buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(java.io.InputStream input)
throws java.io.IOException {
return newBuilder().mergeFrom(input).buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return newBuilder().mergeFrom(input, extensionRegistry)
.buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
Builder builder = newBuilder();
if (builder.mergeDelimitedFrom(input)) {
return builder.buildParsed();
} else {
return null;
}
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
Builder builder = newBuilder();
if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
return builder.buildParsed();
} else {
return null;
}
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return newBuilder().mergeFrom(input).buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return newBuilder().mergeFrom(input, extensionRegistry)
.buildParsed();
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder<Builder> {
private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node result;
// Construct using org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.newBuilder()
private Builder() {}
private static Builder create() {
Builder builder = new Builder();
builder.result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node();
return builder;
}
protected org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node internalGetResult() {
return result;
}
public Builder clear() {
if (result == null) {
throw new IllegalStateException(
"Cannot call clear() after build().");
}
result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node();
return this;
}
public Builder clone() {
return create().mergeFrom(result);
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.getDescriptor();
}
public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node getDefaultInstanceForType() {
return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.getDefaultInstance();
}
public boolean isInitialized() {
return result.isInitialized();
}
public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node build() {
if (result != null && !isInitialized()) {
throw newUninitializedMessageException(result);
}
return buildPartial();
}
private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node buildParsed()
throws com.google.protobuf.InvalidProtocolBufferException {
if (!isInitialized()) {
throw newUninitializedMessageException(
result).asInvalidProtocolBufferException();
}
return buildPartial();
}
public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node buildPartial() {
if (result == null) {
throw new IllegalStateException(
"build() has already been called on this Builder.");
}
if (result.regions_ != java.util.Collections.EMPTY_LIST) {
result.regions_ =
java.util.Collections.unmodifiableList(result.regions_);
}
org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node returnMe = result;
result = null;
return returnMe;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node) {
return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node other) {
if (other == org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.getDefaultInstance()) return this;
if (other.hasName()) {
setName(other.getName());
}
if (other.hasStartCode()) {
setStartCode(other.getStartCode());
}
if (other.hasRequests()) {
setRequests(other.getRequests());
}
if (other.hasHeapSizeMB()) {
setHeapSizeMB(other.getHeapSizeMB());
}
if (other.hasMaxHeapSizeMB()) {
setMaxHeapSizeMB(other.getMaxHeapSizeMB());
}
if (!other.regions_.isEmpty()) {
if (result.regions_.isEmpty()) {
result.regions_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region>();
}
result.regions_.addAll(other.regions_);
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder(
this.getUnknownFields());
while (true) {
int tag = input.readTag();
switch (tag) {
case 0:
this.setUnknownFields(unknownFields.build());
return this;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
this.setUnknownFields(unknownFields.build());
return this;
}
break;
}
case 10: {
setName(input.readString());
break;
}
case 16: {
setStartCode(input.readInt64());
break;
}
case 24: {
setRequests(input.readInt32());
break;
}
case 32: {
setHeapSizeMB(input.readInt32());
break;
}
case 40: {
setMaxHeapSizeMB(input.readInt32());
break;
}
case 50: {
org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder subBuilder = org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.newBuilder();
input.readMessage(subBuilder, extensionRegistry);
addRegions(subBuilder.buildPartial());
break;
}
}
}
}
// required string name = 1;
public boolean hasName() {
return result.hasName();
}
public java.lang.String getName() {
return result.getName();
}
public Builder setName(java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
result.hasName = true;
result.name_ = value;
return this;
}
public Builder clearName() {
result.hasName = false;
result.name_ = getDefaultInstance().getName();
return this;
}
// optional int64 startCode = 2;
public boolean hasStartCode() {
return result.hasStartCode();
}
public long getStartCode() {
return result.getStartCode();
}
public Builder setStartCode(long value) {
result.hasStartCode = true;
result.startCode_ = value;
return this;
}
public Builder clearStartCode() {
result.hasStartCode = false;
result.startCode_ = 0L;
return this;
}
// optional int32 requests = 3;
public boolean hasRequests() {
return result.hasRequests();
}
public int getRequests() {
return result.getRequests();
}
public Builder setRequests(int value) {
result.hasRequests = true;
result.requests_ = value;
return this;
}
public Builder clearRequests() {
result.hasRequests = false;
result.requests_ = 0;
return this;
}
// optional int32 heapSizeMB = 4;
public boolean hasHeapSizeMB() {
return result.hasHeapSizeMB();
}
public int getHeapSizeMB() {
return result.getHeapSizeMB();
}
public Builder setHeapSizeMB(int value) {
result.hasHeapSizeMB = true;
result.heapSizeMB_ = value;
return this;
}
public Builder clearHeapSizeMB() {
result.hasHeapSizeMB = false;
result.heapSizeMB_ = 0;
return this;
}
// optional int32 maxHeapSizeMB = 5;
public boolean hasMaxHeapSizeMB() {
return result.hasMaxHeapSizeMB();
}
public int getMaxHeapSizeMB() {
return result.getMaxHeapSizeMB();
}
public Builder setMaxHeapSizeMB(int value) {
result.hasMaxHeapSizeMB = true;
result.maxHeapSizeMB_ = value;
return this;
}
public Builder clearMaxHeapSizeMB() {
result.hasMaxHeapSizeMB = false;
result.maxHeapSizeMB_ = 0;
return this;
}
// repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Region regions = 6;
public java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> getRegionsList() {
return java.util.Collections.unmodifiableList(result.regions_);
}
public int getRegionsCount() {
return result.getRegionsCount();
}
public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region getRegions(int index) {
return result.getRegions(index);
}
public Builder setRegions(int index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region value) {
if (value == null) {
throw new NullPointerException();
}
result.regions_.set(index, value);
return this;
}
public Builder setRegions(int index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder builderForValue) {
result.regions_.set(index, builderForValue.build());
return this;
}
public Builder addRegions(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region value) {
if (value == null) {
throw new NullPointerException();
}
if (result.regions_.isEmpty()) {
result.regions_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region>();
}
result.regions_.add(value);
return this;
}
public Builder addRegions(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder builderForValue) {
if (result.regions_.isEmpty()) {
result.regions_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region>();
}
result.regions_.add(builderForValue.build());
return this;
}
public Builder addAllRegions(
java.lang.Iterable<? extends org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> values) {
if (result.regions_.isEmpty()) {
result.regions_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region>();
}
super.addAll(values, result.regions_);
return this;
}
public Builder clearRegions() {
result.regions_ = java.util.Collections.emptyList();
return this;
}
// @@protoc_insertion_point(builder_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node)
}
static {
defaultInstance = new Node(true);
org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internalForceInit();
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node)
}
// repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node liveNodes = 1;
public static final int LIVENODES_FIELD_NUMBER = 1;
private java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node> liveNodes_ =
java.util.Collections.emptyList();
public java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node> getLiveNodesList() {
return liveNodes_;
}
public int getLiveNodesCount() { return liveNodes_.size(); }
public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node getLiveNodes(int index) {
return liveNodes_.get(index);
}
// repeated string deadNodes = 2;
public static final int DEADNODES_FIELD_NUMBER = 2;
private java.util.List<java.lang.String> deadNodes_ =
java.util.Collections.emptyList();
public java.util.List<java.lang.String> getDeadNodesList() {
return deadNodes_;
}
public int getDeadNodesCount() { return deadNodes_.size(); }
public java.lang.String getDeadNodes(int index) {
return deadNodes_.get(index);
}
// optional int32 regions = 3;
public static final int REGIONS_FIELD_NUMBER = 3;
private boolean hasRegions;
private int regions_ = 0;
public boolean hasRegions() { return hasRegions; }
public int getRegions() { return regions_; }
// optional int32 requests = 4;
public static final int REQUESTS_FIELD_NUMBER = 4;
private boolean hasRequests;
private int requests_ = 0;
public boolean hasRequests() { return hasRequests; }
public int getRequests() { return requests_; }
// optional double averageLoad = 5;
public static final int AVERAGELOAD_FIELD_NUMBER = 5;
private boolean hasAverageLoad;
private double averageLoad_ = 0D;
public boolean hasAverageLoad() { return hasAverageLoad; }
public double getAverageLoad() { return averageLoad_; }
private void initFields() {
}
public final boolean isInitialized() {
for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node element : getLiveNodesList()) {
if (!element.isInitialized()) return false;
}
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node element : getLiveNodesList()) {
output.writeMessage(1, element);
}
for (java.lang.String element : getDeadNodesList()) {
output.writeString(2, element);
}
if (hasRegions()) {
output.writeInt32(3, getRegions());
}
if (hasRequests()) {
output.writeInt32(4, getRequests());
}
if (hasAverageLoad()) {
output.writeDouble(5, getAverageLoad());
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node element : getLiveNodesList()) {
size += com.google.protobuf.CodedOutputStream
.computeMessageSize(1, element);
}
{
int dataSize = 0;
for (java.lang.String element : getDeadNodesList()) {
dataSize += com.google.protobuf.CodedOutputStream
.computeStringSizeNoTag(element);
}
size += dataSize;
size += 1 * getDeadNodesList().size();
}
if (hasRegions()) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(3, getRegions());
}
if (hasRequests()) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(4, getRequests());
}
if (hasAverageLoad()) {
size += com.google.protobuf.CodedOutputStream
.computeDoubleSize(5, getAverageLoad());
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data).buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data, extensionRegistry)
.buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data).buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data, extensionRegistry)
.buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(java.io.InputStream input)
throws java.io.IOException {
return newBuilder().mergeFrom(input).buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return newBuilder().mergeFrom(input, extensionRegistry)
.buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
Builder builder = newBuilder();
if (builder.mergeDelimitedFrom(input)) {
return builder.buildParsed();
} else {
return null;
}
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
Builder builder = newBuilder();
if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
return builder.buildParsed();
} else {
return null;
}
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return newBuilder().mergeFrom(input).buildParsed();
}
public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return newBuilder().mergeFrom(input, extensionRegistry)
.buildParsed();
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder<Builder> {
private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus result;
// Construct using org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.newBuilder()
private Builder() {}
private static Builder create() {
Builder builder = new Builder();
builder.result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus();
return builder;
}
protected org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus internalGetResult() {
return result;
}
public Builder clear() {
if (result == null) {
throw new IllegalStateException(
"Cannot call clear() after build().");
}
result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus();
return this;
}
public Builder clone() {
return create().mergeFrom(result);
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.getDescriptor();
}
public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus getDefaultInstanceForType() {
return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.getDefaultInstance();
}
public boolean isInitialized() {
return result.isInitialized();
}
public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus build() {
if (result != null && !isInitialized()) {
throw newUninitializedMessageException(result);
}
return buildPartial();
}
private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus buildParsed()
throws com.google.protobuf.InvalidProtocolBufferException {
if (!isInitialized()) {
throw newUninitializedMessageException(
result).asInvalidProtocolBufferException();
}
return buildPartial();
}
public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus buildPartial() {
if (result == null) {
throw new IllegalStateException(
"build() has already been called on this Builder.");
}
if (result.liveNodes_ != java.util.Collections.EMPTY_LIST) {
result.liveNodes_ =
java.util.Collections.unmodifiableList(result.liveNodes_);
}
if (result.deadNodes_ != java.util.Collections.EMPTY_LIST) {
result.deadNodes_ =
java.util.Collections.unmodifiableList(result.deadNodes_);
}
org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus returnMe = result;
result = null;
return returnMe;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus) {
return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus other) {
if (other == org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.getDefaultInstance()) return this;
if (!other.liveNodes_.isEmpty()) {
if (result.liveNodes_.isEmpty()) {
result.liveNodes_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node>();
}
result.liveNodes_.addAll(other.liveNodes_);
}
if (!other.deadNodes_.isEmpty()) {
if (result.deadNodes_.isEmpty()) {
result.deadNodes_ = new java.util.ArrayList<java.lang.String>();
}
result.deadNodes_.addAll(other.deadNodes_);
}
if (other.hasRegions()) {
setRegions(other.getRegions());
}
if (other.hasRequests()) {
setRequests(other.getRequests());
}
if (other.hasAverageLoad()) {
setAverageLoad(other.getAverageLoad());
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder(
this.getUnknownFields());
while (true) {
int tag = input.readTag();
switch (tag) {
case 0:
this.setUnknownFields(unknownFields.build());
return this;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
this.setUnknownFields(unknownFields.build());
return this;
}
break;
}
case 10: {
org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.Builder subBuilder = org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.newBuilder();
input.readMessage(subBuilder, extensionRegistry);
addLiveNodes(subBuilder.buildPartial());
break;
}
case 18: {
addDeadNodes(input.readString());
break;
}
case 24: {
setRegions(input.readInt32());
break;
}
case 32: {
setRequests(input.readInt32());
break;
}
case 41: {
setAverageLoad(input.readDouble());
break;
}
}
}
}
// repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node liveNodes = 1;
public java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node> getLiveNodesList() {
return java.util.Collections.unmodifiableList(result.liveNodes_);
}
public int getLiveNodesCount() {
return result.getLiveNodesCount();
}
public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node getLiveNodes(int index) {
return result.getLiveNodes(index);
}
public Builder setLiveNodes(int index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node value) {
if (value == null) {
throw new NullPointerException();
}
result.liveNodes_.set(index, value);
return this;
}
public Builder setLiveNodes(int index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.Builder builderForValue) {
result.liveNodes_.set(index, builderForValue.build());
return this;
}
public Builder addLiveNodes(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node value) {
if (value == null) {
throw new NullPointerException();
}
if (result.liveNodes_.isEmpty()) {
result.liveNodes_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node>();
}
result.liveNodes_.add(value);
return this;
}
public Builder addLiveNodes(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.Builder builderForValue) {
if (result.liveNodes_.isEmpty()) {
result.liveNodes_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node>();
}
result.liveNodes_.add(builderForValue.build());
return this;
}
public Builder addAllLiveNodes(
java.lang.Iterable<? extends org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node> values) {
if (result.liveNodes_.isEmpty()) {
result.liveNodes_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node>();
}
super.addAll(values, result.liveNodes_);
return this;
}
public Builder clearLiveNodes() {
result.liveNodes_ = java.util.Collections.emptyList();
return this;
}
// repeated string deadNodes = 2;
public java.util.List<java.lang.String> getDeadNodesList() {
return java.util.Collections.unmodifiableList(result.deadNodes_);
}
public int getDeadNodesCount() {
return result.getDeadNodesCount();
}
public java.lang.String getDeadNodes(int index) {
return result.getDeadNodes(index);
}
public Builder setDeadNodes(int index, java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
result.deadNodes_.set(index, value);
return this;
}
public Builder addDeadNodes(java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
if (result.deadNodes_.isEmpty()) {
result.deadNodes_ = new java.util.ArrayList<java.lang.String>();
}
result.deadNodes_.add(value);
return this;
}
public Builder addAllDeadNodes(
java.lang.Iterable<? extends java.lang.String> values) {
if (result.deadNodes_.isEmpty()) {
result.deadNodes_ = new java.util.ArrayList<java.lang.String>();
}
super.addAll(values, result.deadNodes_);
return this;
}
public Builder clearDeadNodes() {
result.deadNodes_ = java.util.Collections.emptyList();
return this;
}
// optional int32 regions = 3;
public boolean hasRegions() {
return result.hasRegions();
}
public int getRegions() {
return result.getRegions();
}
public Builder setRegions(int value) {
result.hasRegions = true;
result.regions_ = value;
return this;
}
public Builder clearRegions() {
result.hasRegions = false;
result.regions_ = 0;
return this;
}
// optional int32 requests = 4;
public boolean hasRequests() {
return result.hasRequests();
}
public int getRequests() {
return result.getRequests();
}
public Builder setRequests(int value) {
result.hasRequests = true;
result.requests_ = value;
return this;
}
public Builder clearRequests() {
result.hasRequests = false;
result.requests_ = 0;
return this;
}
// optional double averageLoad = 5;
public boolean hasAverageLoad() {
return result.hasAverageLoad();
}
public double getAverageLoad() {
return result.getAverageLoad();
}
public Builder setAverageLoad(double value) {
result.hasAverageLoad = true;
result.averageLoad_ = value;
return this;
}
public Builder clearAverageLoad() {
result.hasAverageLoad = false;
result.averageLoad_ = 0D;
return this;
}
// @@protoc_insertion_point(builder_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus)
}
static {
defaultInstance = new StorageClusterStatus(true);
org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internalForceInit();
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus)
}
private static com.google.protobuf.Descriptors.Descriptor
internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_fieldAccessorTable;
public static com.google.protobuf.Descriptors.FileDescriptor
getDescriptor() {
return descriptor;
}
private static com.google.protobuf.Descriptors.FileDescriptor
descriptor;
static {
java.lang.String[] descriptorData = {
"\n!StorageClusterStatusMessage.proto\022/org" +
".apache.hadoop.hbase.rest.protobuf.gener" +
"ated\"\222\004\n\024StorageClusterStatus\022]\n\tliveNod" +
"es\030\001 \003(\0132J.org.apache.hadoop.hbase.rest." +
"protobuf.generated.StorageClusterStatus." +
"Node\022\021\n\tdeadNodes\030\002 \003(\t\022\017\n\007regions\030\003 \001(\005" +
"\022\020\n\010requests\030\004 \001(\005\022\023\n\013averageLoad\030\005 \001(\001\032" +
"\211\001\n\006Region\022\014\n\004name\030\001 \002(\014\022\016\n\006stores\030\002 \001(\005" +
"\022\022\n\nstorefiles\030\003 \001(\005\022\027\n\017storefileSizeMB\030" +
"\004 \001(\005\022\026\n\016memstoreSizeMB\030\005 \001(\005\022\034\n\024storefi",
"leIndexSizeMB\030\006 \001(\005\032\303\001\n\004Node\022\014\n\004name\030\001 \002" +
"(\t\022\021\n\tstartCode\030\002 \001(\003\022\020\n\010requests\030\003 \001(\005\022" +
"\022\n\nheapSizeMB\030\004 \001(\005\022\025\n\rmaxHeapSizeMB\030\005 \001" +
"(\005\022]\n\007regions\030\006 \003(\0132L.org.apache.hadoop." +
"hbase.rest.protobuf.generated.StorageClu" +
"sterStatus.Region"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
public com.google.protobuf.ExtensionRegistry assignDescriptors(
com.google.protobuf.Descriptors.FileDescriptor root) {
descriptor = root;
internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor =
getDescriptor().getMessageTypes().get(0);
internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor,
new java.lang.String[] { "LiveNodes", "DeadNodes", "Regions", "Requests", "AverageLoad", },
org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.class,
org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Builder.class);
internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor =
internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor.getNestedTypes().get(0);
internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor,
new java.lang.String[] { "Name", "Stores", "Storefiles", "StorefileSizeMB", "MemstoreSizeMB", "StorefileIndexSizeMB", },
org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.class,
org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder.class);
internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor =
internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor.getNestedTypes().get(1);
internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor,
new java.lang.String[] { "Name", "StartCode", "Requests", "HeapSizeMB", "MaxHeapSizeMB", "Regions", },
org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.class,
org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.Builder.class);
return null;
}
};
com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
new com.google.protobuf.Descriptors.FileDescriptor[] {
}, assigner);
}
public static void internalForceInit() {}
// @@protoc_insertion_point(outer_class_scope)
}