// Generated by the protocol buffer compiler. DO NOT EDIT! // source: StorageClusterStatusMessage.proto package org.apache.hadoop.hbase.rest.protobuf.generated; public final class StorageClusterStatusMessage { private StorageClusterStatusMessage() {} public static void registerAllExtensions( com.google.protobuf.ExtensionRegistry registry) { } public static final class StorageClusterStatus extends com.google.protobuf.GeneratedMessage { // Use StorageClusterStatus.newBuilder() to construct. private StorageClusterStatus() { initFields(); } private StorageClusterStatus(boolean noInit) {} private static final StorageClusterStatus defaultInstance; public static StorageClusterStatus getDefaultInstance() { return defaultInstance; } public StorageClusterStatus getDefaultInstanceForType() { return defaultInstance; } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_fieldAccessorTable; } public static final class Region extends com.google.protobuf.GeneratedMessage { // Use Region.newBuilder() to construct. private Region() { initFields(); } private Region(boolean noInit) {} private static final Region defaultInstance; public static Region getDefaultInstance() { return defaultInstance; } public Region getDefaultInstanceForType() { return defaultInstance; } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_fieldAccessorTable; } // required bytes name = 1; public static final int NAME_FIELD_NUMBER = 1; private boolean hasName; private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY; public boolean hasName() { return hasName; } public com.google.protobuf.ByteString getName() { return name_; } // optional int32 stores = 2; public static final int STORES_FIELD_NUMBER = 2; private boolean hasStores; private int stores_ = 0; public boolean hasStores() { return hasStores; } public int getStores() { return stores_; } // optional int32 storefiles = 3; public static final int STOREFILES_FIELD_NUMBER = 3; private boolean hasStorefiles; private int storefiles_ = 0; public boolean hasStorefiles() { return hasStorefiles; } public int getStorefiles() { return storefiles_; } // optional int32 storefileSizeMB = 4; public static final int STOREFILESIZEMB_FIELD_NUMBER = 4; private boolean hasStorefileSizeMB; private int storefileSizeMB_ = 0; public boolean hasStorefileSizeMB() { return hasStorefileSizeMB; } public int getStorefileSizeMB() { return storefileSizeMB_; } // optional int32 memstoreSizeMB = 5; public static final int MEMSTORESIZEMB_FIELD_NUMBER = 5; private boolean hasMemstoreSizeMB; private int memstoreSizeMB_ = 0; public boolean hasMemstoreSizeMB() { return hasMemstoreSizeMB; } public int getMemstoreSizeMB() { return memstoreSizeMB_; } // optional int32 storefileIndexSizeMB = 6; public static final int STOREFILEINDEXSIZEMB_FIELD_NUMBER = 6; private boolean hasStorefileIndexSizeMB; private int storefileIndexSizeMB_ = 0; public boolean hasStorefileIndexSizeMB() { return hasStorefileIndexSizeMB; } public int getStorefileIndexSizeMB() { return storefileIndexSizeMB_; } private void initFields() { } public final boolean isInitialized() { if (!hasName) return false; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (hasName()) { output.writeBytes(1, getName()); } if (hasStores()) { output.writeInt32(2, getStores()); } if (hasStorefiles()) { output.writeInt32(3, getStorefiles()); } if (hasStorefileSizeMB()) { output.writeInt32(4, getStorefileSizeMB()); } if (hasMemstoreSizeMB()) { output.writeInt32(5, getMemstoreSizeMB()); } if (hasStorefileIndexSizeMB()) { output.writeInt32(6, getStorefileIndexSizeMB()); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (hasName()) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, getName()); } if (hasStores()) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(2, getStores()); } if (hasStorefiles()) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(3, getStorefiles()); } if (hasStorefileSizeMB()) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(4, getStorefileSizeMB()); } if (hasMemstoreSizeMB()) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(5, getMemstoreSizeMB()); } if (hasStorefileIndexSizeMB()) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(6, getStorefileIndexSizeMB()); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(java.io.InputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return newBuilder().mergeFrom(input, extensionRegistry) .buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { Builder builder = newBuilder(); if (builder.mergeDelimitedFrom(input)) { return builder.buildParsed(); } else { return null; } } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { Builder builder = newBuilder(); if (builder.mergeDelimitedFrom(input, extensionRegistry)) { return builder.buildParsed(); } else { return null; } } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return newBuilder().mergeFrom(input, extensionRegistry) .buildParsed(); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> { private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region result; // Construct using org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.newBuilder() private Builder() {} private static Builder create() { Builder builder = new Builder(); builder.result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region(); return builder; } protected org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region internalGetResult() { return result; } public Builder clear() { if (result == null) { throw new IllegalStateException( "Cannot call clear() after build()."); } result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region(); return this; } public Builder clone() { return create().mergeFrom(result); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.getDescriptor(); } public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region getDefaultInstanceForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.getDefaultInstance(); } public boolean isInitialized() { return result.isInitialized(); } public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region build() { if (result != null && !isInitialized()) { throw newUninitializedMessageException(result); } return buildPartial(); } private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region buildParsed() throws com.google.protobuf.InvalidProtocolBufferException { if (!isInitialized()) { throw newUninitializedMessageException( result).asInvalidProtocolBufferException(); } return buildPartial(); } public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region buildPartial() { if (result == null) { throw new IllegalStateException( "build() has already been called on this Builder."); } org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region returnMe = result; result = null; return returnMe; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region other) { if (other == org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.getDefaultInstance()) return this; if (other.hasName()) { setName(other.getName()); } if (other.hasStores()) { setStores(other.getStores()); } if (other.hasStorefiles()) { setStorefiles(other.getStorefiles()); } if (other.hasStorefileSizeMB()) { setStorefileSizeMB(other.getStorefileSizeMB()); } if (other.hasMemstoreSizeMB()) { setMemstoreSizeMB(other.getMemstoreSizeMB()); } if (other.hasStorefileIndexSizeMB()) { setStorefileIndexSizeMB(other.getStorefileIndexSizeMB()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder( this.getUnknownFields()); while (true) { int tag = input.readTag(); switch (tag) { case 0: this.setUnknownFields(unknownFields.build()); return this; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { this.setUnknownFields(unknownFields.build()); return this; } break; } case 10: { setName(input.readBytes()); break; } case 16: { setStores(input.readInt32()); break; } case 24: { setStorefiles(input.readInt32()); break; } case 32: { setStorefileSizeMB(input.readInt32()); break; } case 40: { setMemstoreSizeMB(input.readInt32()); break; } case 48: { setStorefileIndexSizeMB(input.readInt32()); break; } } } } // required bytes name = 1; public boolean hasName() { return result.hasName(); } public com.google.protobuf.ByteString getName() { return result.getName(); } public Builder setName(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } result.hasName = true; result.name_ = value; return this; } public Builder clearName() { result.hasName = false; result.name_ = getDefaultInstance().getName(); return this; } // optional int32 stores = 2; public boolean hasStores() { return result.hasStores(); } public int getStores() { return result.getStores(); } public Builder setStores(int value) { result.hasStores = true; result.stores_ = value; return this; } public Builder clearStores() { result.hasStores = false; result.stores_ = 0; return this; } // optional int32 storefiles = 3; public boolean hasStorefiles() { return result.hasStorefiles(); } public int getStorefiles() { return result.getStorefiles(); } public Builder setStorefiles(int value) { result.hasStorefiles = true; result.storefiles_ = value; return this; } public Builder clearStorefiles() { result.hasStorefiles = false; result.storefiles_ = 0; return this; } // optional int32 storefileSizeMB = 4; public boolean hasStorefileSizeMB() { return result.hasStorefileSizeMB(); } public int getStorefileSizeMB() { return result.getStorefileSizeMB(); } public Builder setStorefileSizeMB(int value) { result.hasStorefileSizeMB = true; result.storefileSizeMB_ = value; return this; } public Builder clearStorefileSizeMB() { result.hasStorefileSizeMB = false; result.storefileSizeMB_ = 0; return this; } // optional int32 memstoreSizeMB = 5; public boolean hasMemstoreSizeMB() { return result.hasMemstoreSizeMB(); } public int getMemstoreSizeMB() { return result.getMemstoreSizeMB(); } public Builder setMemstoreSizeMB(int value) { result.hasMemstoreSizeMB = true; result.memstoreSizeMB_ = value; return this; } public Builder clearMemstoreSizeMB() { result.hasMemstoreSizeMB = false; result.memstoreSizeMB_ = 0; return this; } // optional int32 storefileIndexSizeMB = 6; public boolean hasStorefileIndexSizeMB() { return result.hasStorefileIndexSizeMB(); } public int getStorefileIndexSizeMB() { return result.getStorefileIndexSizeMB(); } public Builder setStorefileIndexSizeMB(int value) { result.hasStorefileIndexSizeMB = true; result.storefileIndexSizeMB_ = value; return this; } public Builder clearStorefileIndexSizeMB() { result.hasStorefileIndexSizeMB = false; result.storefileIndexSizeMB_ = 0; return this; } // @@protoc_insertion_point(builder_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Region) } static { defaultInstance = new Region(true); org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internalForceInit(); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Region) } public static final class Node extends com.google.protobuf.GeneratedMessage { // Use Node.newBuilder() to construct. private Node() { initFields(); } private Node(boolean noInit) {} private static final Node defaultInstance; public static Node getDefaultInstance() { return defaultInstance; } public Node getDefaultInstanceForType() { return defaultInstance; } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_fieldAccessorTable; } // required string name = 1; public static final int NAME_FIELD_NUMBER = 1; private boolean hasName; private java.lang.String name_ = ""; public boolean hasName() { return hasName; } public java.lang.String getName() { return name_; } // optional int64 startCode = 2; public static final int STARTCODE_FIELD_NUMBER = 2; private boolean hasStartCode; private long startCode_ = 0L; public boolean hasStartCode() { return hasStartCode; } public long getStartCode() { return startCode_; } // optional int32 requests = 3; public static final int REQUESTS_FIELD_NUMBER = 3; private boolean hasRequests; private int requests_ = 0; public boolean hasRequests() { return hasRequests; } public int getRequests() { return requests_; } // optional int32 heapSizeMB = 4; public static final int HEAPSIZEMB_FIELD_NUMBER = 4; private boolean hasHeapSizeMB; private int heapSizeMB_ = 0; public boolean hasHeapSizeMB() { return hasHeapSizeMB; } public int getHeapSizeMB() { return heapSizeMB_; } // optional int32 maxHeapSizeMB = 5; public static final int MAXHEAPSIZEMB_FIELD_NUMBER = 5; private boolean hasMaxHeapSizeMB; private int maxHeapSizeMB_ = 0; public boolean hasMaxHeapSizeMB() { return hasMaxHeapSizeMB; } public int getMaxHeapSizeMB() { return maxHeapSizeMB_; } // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Region regions = 6; public static final int REGIONS_FIELD_NUMBER = 6; private java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> regions_ = java.util.Collections.emptyList(); public java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> getRegionsList() { return regions_; } public int getRegionsCount() { return regions_.size(); } public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region getRegions(int index) { return regions_.get(index); } private void initFields() { } public final boolean isInitialized() { if (!hasName) return false; for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region element : getRegionsList()) { if (!element.isInitialized()) return false; } return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (hasName()) { output.writeString(1, getName()); } if (hasStartCode()) { output.writeInt64(2, getStartCode()); } if (hasRequests()) { output.writeInt32(3, getRequests()); } if (hasHeapSizeMB()) { output.writeInt32(4, getHeapSizeMB()); } if (hasMaxHeapSizeMB()) { output.writeInt32(5, getMaxHeapSizeMB()); } for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region element : getRegionsList()) { output.writeMessage(6, element); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (hasName()) { size += com.google.protobuf.CodedOutputStream .computeStringSize(1, getName()); } if (hasStartCode()) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(2, getStartCode()); } if (hasRequests()) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(3, getRequests()); } if (hasHeapSizeMB()) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(4, getHeapSizeMB()); } if (hasMaxHeapSizeMB()) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(5, getMaxHeapSizeMB()); } for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region element : getRegionsList()) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(6, element); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(java.io.InputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return newBuilder().mergeFrom(input, extensionRegistry) .buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { Builder builder = newBuilder(); if (builder.mergeDelimitedFrom(input)) { return builder.buildParsed(); } else { return null; } } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { Builder builder = newBuilder(); if (builder.mergeDelimitedFrom(input, extensionRegistry)) { return builder.buildParsed(); } else { return null; } } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return newBuilder().mergeFrom(input, extensionRegistry) .buildParsed(); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> { private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node result; // Construct using org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.newBuilder() private Builder() {} private static Builder create() { Builder builder = new Builder(); builder.result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node(); return builder; } protected org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node internalGetResult() { return result; } public Builder clear() { if (result == null) { throw new IllegalStateException( "Cannot call clear() after build()."); } result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node(); return this; } public Builder clone() { return create().mergeFrom(result); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.getDescriptor(); } public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node getDefaultInstanceForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.getDefaultInstance(); } public boolean isInitialized() { return result.isInitialized(); } public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node build() { if (result != null && !isInitialized()) { throw newUninitializedMessageException(result); } return buildPartial(); } private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node buildParsed() throws com.google.protobuf.InvalidProtocolBufferException { if (!isInitialized()) { throw newUninitializedMessageException( result).asInvalidProtocolBufferException(); } return buildPartial(); } public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node buildPartial() { if (result == null) { throw new IllegalStateException( "build() has already been called on this Builder."); } if (result.regions_ != java.util.Collections.EMPTY_LIST) { result.regions_ = java.util.Collections.unmodifiableList(result.regions_); } org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node returnMe = result; result = null; return returnMe; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node other) { if (other == org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.getDefaultInstance()) return this; if (other.hasName()) { setName(other.getName()); } if (other.hasStartCode()) { setStartCode(other.getStartCode()); } if (other.hasRequests()) { setRequests(other.getRequests()); } if (other.hasHeapSizeMB()) { setHeapSizeMB(other.getHeapSizeMB()); } if (other.hasMaxHeapSizeMB()) { setMaxHeapSizeMB(other.getMaxHeapSizeMB()); } if (!other.regions_.isEmpty()) { if (result.regions_.isEmpty()) { result.regions_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region>(); } result.regions_.addAll(other.regions_); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder( this.getUnknownFields()); while (true) { int tag = input.readTag(); switch (tag) { case 0: this.setUnknownFields(unknownFields.build()); return this; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { this.setUnknownFields(unknownFields.build()); return this; } break; } case 10: { setName(input.readString()); break; } case 16: { setStartCode(input.readInt64()); break; } case 24: { setRequests(input.readInt32()); break; } case 32: { setHeapSizeMB(input.readInt32()); break; } case 40: { setMaxHeapSizeMB(input.readInt32()); break; } case 50: { org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder subBuilder = org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.newBuilder(); input.readMessage(subBuilder, extensionRegistry); addRegions(subBuilder.buildPartial()); break; } } } } // required string name = 1; public boolean hasName() { return result.hasName(); } public java.lang.String getName() { return result.getName(); } public Builder setName(java.lang.String value) { if (value == null) { throw new NullPointerException(); } result.hasName = true; result.name_ = value; return this; } public Builder clearName() { result.hasName = false; result.name_ = getDefaultInstance().getName(); return this; } // optional int64 startCode = 2; public boolean hasStartCode() { return result.hasStartCode(); } public long getStartCode() { return result.getStartCode(); } public Builder setStartCode(long value) { result.hasStartCode = true; result.startCode_ = value; return this; } public Builder clearStartCode() { result.hasStartCode = false; result.startCode_ = 0L; return this; } // optional int32 requests = 3; public boolean hasRequests() { return result.hasRequests(); } public int getRequests() { return result.getRequests(); } public Builder setRequests(int value) { result.hasRequests = true; result.requests_ = value; return this; } public Builder clearRequests() { result.hasRequests = false; result.requests_ = 0; return this; } // optional int32 heapSizeMB = 4; public boolean hasHeapSizeMB() { return result.hasHeapSizeMB(); } public int getHeapSizeMB() { return result.getHeapSizeMB(); } public Builder setHeapSizeMB(int value) { result.hasHeapSizeMB = true; result.heapSizeMB_ = value; return this; } public Builder clearHeapSizeMB() { result.hasHeapSizeMB = false; result.heapSizeMB_ = 0; return this; } // optional int32 maxHeapSizeMB = 5; public boolean hasMaxHeapSizeMB() { return result.hasMaxHeapSizeMB(); } public int getMaxHeapSizeMB() { return result.getMaxHeapSizeMB(); } public Builder setMaxHeapSizeMB(int value) { result.hasMaxHeapSizeMB = true; result.maxHeapSizeMB_ = value; return this; } public Builder clearMaxHeapSizeMB() { result.hasMaxHeapSizeMB = false; result.maxHeapSizeMB_ = 0; return this; } // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Region regions = 6; public java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> getRegionsList() { return java.util.Collections.unmodifiableList(result.regions_); } public int getRegionsCount() { return result.getRegionsCount(); } public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region getRegions(int index) { return result.getRegions(index); } public Builder setRegions(int index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region value) { if (value == null) { throw new NullPointerException(); } result.regions_.set(index, value); return this; } public Builder setRegions(int index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder builderForValue) { result.regions_.set(index, builderForValue.build()); return this; } public Builder addRegions(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region value) { if (value == null) { throw new NullPointerException(); } if (result.regions_.isEmpty()) { result.regions_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region>(); } result.regions_.add(value); return this; } public Builder addRegions(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder builderForValue) { if (result.regions_.isEmpty()) { result.regions_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region>(); } result.regions_.add(builderForValue.build()); return this; } public Builder addAllRegions( java.lang.Iterable<? extends org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> values) { if (result.regions_.isEmpty()) { result.regions_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region>(); } super.addAll(values, result.regions_); return this; } public Builder clearRegions() { result.regions_ = java.util.Collections.emptyList(); return this; } // @@protoc_insertion_point(builder_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node) } static { defaultInstance = new Node(true); org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internalForceInit(); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node) } // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node liveNodes = 1; public static final int LIVENODES_FIELD_NUMBER = 1; private java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node> liveNodes_ = java.util.Collections.emptyList(); public java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node> getLiveNodesList() { return liveNodes_; } public int getLiveNodesCount() { return liveNodes_.size(); } public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node getLiveNodes(int index) { return liveNodes_.get(index); } // repeated string deadNodes = 2; public static final int DEADNODES_FIELD_NUMBER = 2; private java.util.List<java.lang.String> deadNodes_ = java.util.Collections.emptyList(); public java.util.List<java.lang.String> getDeadNodesList() { return deadNodes_; } public int getDeadNodesCount() { return deadNodes_.size(); } public java.lang.String getDeadNodes(int index) { return deadNodes_.get(index); } // optional int32 regions = 3; public static final int REGIONS_FIELD_NUMBER = 3; private boolean hasRegions; private int regions_ = 0; public boolean hasRegions() { return hasRegions; } public int getRegions() { return regions_; } // optional int32 requests = 4; public static final int REQUESTS_FIELD_NUMBER = 4; private boolean hasRequests; private int requests_ = 0; public boolean hasRequests() { return hasRequests; } public int getRequests() { return requests_; } // optional double averageLoad = 5; public static final int AVERAGELOAD_FIELD_NUMBER = 5; private boolean hasAverageLoad; private double averageLoad_ = 0D; public boolean hasAverageLoad() { return hasAverageLoad; } public double getAverageLoad() { return averageLoad_; } private void initFields() { } public final boolean isInitialized() { for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node element : getLiveNodesList()) { if (!element.isInitialized()) return false; } return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node element : getLiveNodesList()) { output.writeMessage(1, element); } for (java.lang.String element : getDeadNodesList()) { output.writeString(2, element); } if (hasRegions()) { output.writeInt32(3, getRegions()); } if (hasRequests()) { output.writeInt32(4, getRequests()); } if (hasAverageLoad()) { output.writeDouble(5, getAverageLoad()); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node element : getLiveNodesList()) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(1, element); } { int dataSize = 0; for (java.lang.String element : getDeadNodesList()) { dataSize += com.google.protobuf.CodedOutputStream .computeStringSizeNoTag(element); } size += dataSize; size += 1 * getDeadNodesList().size(); } if (hasRegions()) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(3, getRegions()); } if (hasRequests()) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(4, getRequests()); } if (hasAverageLoad()) { size += com.google.protobuf.CodedOutputStream .computeDoubleSize(5, getAverageLoad()); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data).buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return newBuilder().mergeFrom(data, extensionRegistry) .buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(java.io.InputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return newBuilder().mergeFrom(input, extensionRegistry) .buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { Builder builder = newBuilder(); if (builder.mergeDelimitedFrom(input)) { return builder.buildParsed(); } else { return null; } } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { Builder builder = newBuilder(); if (builder.mergeDelimitedFrom(input, extensionRegistry)) { return builder.buildParsed(); } else { return null; } } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return newBuilder().mergeFrom(input).buildParsed(); } public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return newBuilder().mergeFrom(input, extensionRegistry) .buildParsed(); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> { private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus result; // Construct using org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.newBuilder() private Builder() {} private static Builder create() { Builder builder = new Builder(); builder.result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus(); return builder; } protected org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus internalGetResult() { return result; } public Builder clear() { if (result == null) { throw new IllegalStateException( "Cannot call clear() after build()."); } result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus(); return this; } public Builder clone() { return create().mergeFrom(result); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.getDescriptor(); } public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus getDefaultInstanceForType() { return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.getDefaultInstance(); } public boolean isInitialized() { return result.isInitialized(); } public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus build() { if (result != null && !isInitialized()) { throw newUninitializedMessageException(result); } return buildPartial(); } private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus buildParsed() throws com.google.protobuf.InvalidProtocolBufferException { if (!isInitialized()) { throw newUninitializedMessageException( result).asInvalidProtocolBufferException(); } return buildPartial(); } public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus buildPartial() { if (result == null) { throw new IllegalStateException( "build() has already been called on this Builder."); } if (result.liveNodes_ != java.util.Collections.EMPTY_LIST) { result.liveNodes_ = java.util.Collections.unmodifiableList(result.liveNodes_); } if (result.deadNodes_ != java.util.Collections.EMPTY_LIST) { result.deadNodes_ = java.util.Collections.unmodifiableList(result.deadNodes_); } org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus returnMe = result; result = null; return returnMe; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus) { return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus other) { if (other == org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.getDefaultInstance()) return this; if (!other.liveNodes_.isEmpty()) { if (result.liveNodes_.isEmpty()) { result.liveNodes_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node>(); } result.liveNodes_.addAll(other.liveNodes_); } if (!other.deadNodes_.isEmpty()) { if (result.deadNodes_.isEmpty()) { result.deadNodes_ = new java.util.ArrayList<java.lang.String>(); } result.deadNodes_.addAll(other.deadNodes_); } if (other.hasRegions()) { setRegions(other.getRegions()); } if (other.hasRequests()) { setRequests(other.getRequests()); } if (other.hasAverageLoad()) { setAverageLoad(other.getAverageLoad()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder( this.getUnknownFields()); while (true) { int tag = input.readTag(); switch (tag) { case 0: this.setUnknownFields(unknownFields.build()); return this; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { this.setUnknownFields(unknownFields.build()); return this; } break; } case 10: { org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.Builder subBuilder = org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.newBuilder(); input.readMessage(subBuilder, extensionRegistry); addLiveNodes(subBuilder.buildPartial()); break; } case 18: { addDeadNodes(input.readString()); break; } case 24: { setRegions(input.readInt32()); break; } case 32: { setRequests(input.readInt32()); break; } case 41: { setAverageLoad(input.readDouble()); break; } } } } // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node liveNodes = 1; public java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node> getLiveNodesList() { return java.util.Collections.unmodifiableList(result.liveNodes_); } public int getLiveNodesCount() { return result.getLiveNodesCount(); } public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node getLiveNodes(int index) { return result.getLiveNodes(index); } public Builder setLiveNodes(int index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node value) { if (value == null) { throw new NullPointerException(); } result.liveNodes_.set(index, value); return this; } public Builder setLiveNodes(int index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.Builder builderForValue) { result.liveNodes_.set(index, builderForValue.build()); return this; } public Builder addLiveNodes(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node value) { if (value == null) { throw new NullPointerException(); } if (result.liveNodes_.isEmpty()) { result.liveNodes_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node>(); } result.liveNodes_.add(value); return this; } public Builder addLiveNodes(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.Builder builderForValue) { if (result.liveNodes_.isEmpty()) { result.liveNodes_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node>(); } result.liveNodes_.add(builderForValue.build()); return this; } public Builder addAllLiveNodes( java.lang.Iterable<? extends org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node> values) { if (result.liveNodes_.isEmpty()) { result.liveNodes_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node>(); } super.addAll(values, result.liveNodes_); return this; } public Builder clearLiveNodes() { result.liveNodes_ = java.util.Collections.emptyList(); return this; } // repeated string deadNodes = 2; public java.util.List<java.lang.String> getDeadNodesList() { return java.util.Collections.unmodifiableList(result.deadNodes_); } public int getDeadNodesCount() { return result.getDeadNodesCount(); } public java.lang.String getDeadNodes(int index) { return result.getDeadNodes(index); } public Builder setDeadNodes(int index, java.lang.String value) { if (value == null) { throw new NullPointerException(); } result.deadNodes_.set(index, value); return this; } public Builder addDeadNodes(java.lang.String value) { if (value == null) { throw new NullPointerException(); } if (result.deadNodes_.isEmpty()) { result.deadNodes_ = new java.util.ArrayList<java.lang.String>(); } result.deadNodes_.add(value); return this; } public Builder addAllDeadNodes( java.lang.Iterable<? extends java.lang.String> values) { if (result.deadNodes_.isEmpty()) { result.deadNodes_ = new java.util.ArrayList<java.lang.String>(); } super.addAll(values, result.deadNodes_); return this; } public Builder clearDeadNodes() { result.deadNodes_ = java.util.Collections.emptyList(); return this; } // optional int32 regions = 3; public boolean hasRegions() { return result.hasRegions(); } public int getRegions() { return result.getRegions(); } public Builder setRegions(int value) { result.hasRegions = true; result.regions_ = value; return this; } public Builder clearRegions() { result.hasRegions = false; result.regions_ = 0; return this; } // optional int32 requests = 4; public boolean hasRequests() { return result.hasRequests(); } public int getRequests() { return result.getRequests(); } public Builder setRequests(int value) { result.hasRequests = true; result.requests_ = value; return this; } public Builder clearRequests() { result.hasRequests = false; result.requests_ = 0; return this; } // optional double averageLoad = 5; public boolean hasAverageLoad() { return result.hasAverageLoad(); } public double getAverageLoad() { return result.getAverageLoad(); } public Builder setAverageLoad(double value) { result.hasAverageLoad = true; result.averageLoad_ = value; return this; } public Builder clearAverageLoad() { result.hasAverageLoad = false; result.averageLoad_ = 0D; return this; } // @@protoc_insertion_point(builder_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus) } static { defaultInstance = new StorageClusterStatus(true); org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internalForceInit(); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus) } private static com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_fieldAccessorTable; public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } private static com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { "\n!StorageClusterStatusMessage.proto\022/org" + ".apache.hadoop.hbase.rest.protobuf.gener" + "ated\"\222\004\n\024StorageClusterStatus\022]\n\tliveNod" + "es\030\001 \003(\0132J.org.apache.hadoop.hbase.rest." + "protobuf.generated.StorageClusterStatus." + "Node\022\021\n\tdeadNodes\030\002 \003(\t\022\017\n\007regions\030\003 \001(\005" + "\022\020\n\010requests\030\004 \001(\005\022\023\n\013averageLoad\030\005 \001(\001\032" + "\211\001\n\006Region\022\014\n\004name\030\001 \002(\014\022\016\n\006stores\030\002 \001(\005" + "\022\022\n\nstorefiles\030\003 \001(\005\022\027\n\017storefileSizeMB\030" + "\004 \001(\005\022\026\n\016memstoreSizeMB\030\005 \001(\005\022\034\n\024storefi", "leIndexSizeMB\030\006 \001(\005\032\303\001\n\004Node\022\014\n\004name\030\001 \002" + "(\t\022\021\n\tstartCode\030\002 \001(\003\022\020\n\010requests\030\003 \001(\005\022" + "\022\n\nheapSizeMB\030\004 \001(\005\022\025\n\rmaxHeapSizeMB\030\005 \001" + "(\005\022]\n\007regions\030\006 \003(\0132L.org.apache.hadoop." + "hbase.rest.protobuf.generated.StorageClu" + "sterStatus.Region" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { public com.google.protobuf.ExtensionRegistry assignDescriptors( com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor, new java.lang.String[] { "LiveNodes", "DeadNodes", "Regions", "Requests", "AverageLoad", }, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.class, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Builder.class); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor = internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor.getNestedTypes().get(0); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor, new java.lang.String[] { "Name", "Stores", "Storefiles", "StorefileSizeMB", "MemstoreSizeMB", "StorefileIndexSizeMB", }, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.class, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder.class); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor = internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor.getNestedTypes().get(1); internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor, new java.lang.String[] { "Name", "StartCode", "Requests", "HeapSizeMB", "MaxHeapSizeMB", "Regions", }, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.class, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.Builder.class); return null; } }; com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { }, assigner); } public static void internalForceInit() {} // @@protoc_insertion_point(outer_class_scope) }