hbase-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From li...@apache.org
Subject svn commit: r1594423 [7/17] - in /hbase/branches/0.89-fb: ./ bin/ src/main/java/org/apache/hadoop/hbase/rest/ src/main/java/org/apache/hadoop/hbase/rest/client/ src/main/java/org/apache/hadoop/hbase/rest/metrics/ src/main/java/org/apache/hadoop/hbase/r...
Date Wed, 14 May 2014 00:27:04 GMT
Modified: hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/StorageClusterStatusMessage.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/StorageClusterStatusMessage.java?rev=1594423&r1=1594422&r2=1594423&view=diff
==============================================================================
--- hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/StorageClusterStatusMessage.java (original)
+++ hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/StorageClusterStatusMessage.java Wed May 14 00:26:57 2014
@@ -1,1638 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: StorageClusterStatusMessage.proto
-
-package org.apache.hadoop.hbase.rest.protobuf.generated;
-
-public final class StorageClusterStatusMessage {
-  private StorageClusterStatusMessage() {}
-  public static void registerAllExtensions(
-      com.google.protobuf.ExtensionRegistry registry) {
-  }
-  public static final class StorageClusterStatus extends
-      com.google.protobuf.GeneratedMessage {
-    // Use StorageClusterStatus.newBuilder() to construct.
-    private StorageClusterStatus() {
-      initFields();
-    }
-    private StorageClusterStatus(boolean noInit) {}
-    
-    private static final StorageClusterStatus defaultInstance;
-    public static StorageClusterStatus getDefaultInstance() {
-      return defaultInstance;
-    }
-    
-    public StorageClusterStatus getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-    
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor;
-    }
-    
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_fieldAccessorTable;
-    }
-    
-    public static final class Region extends
-        com.google.protobuf.GeneratedMessage {
-      // Use Region.newBuilder() to construct.
-      private Region() {
-        initFields();
-      }
-      private Region(boolean noInit) {}
-      
-      private static final Region defaultInstance;
-      public static Region getDefaultInstance() {
-        return defaultInstance;
-      }
-      
-      public Region getDefaultInstanceForType() {
-        return defaultInstance;
-      }
-      
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor;
-      }
-      
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_fieldAccessorTable;
-      }
-      
-      // required bytes name = 1;
-      public static final int NAME_FIELD_NUMBER = 1;
-      private boolean hasName;
-      private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY;
-      public boolean hasName() { return hasName; }
-      public com.google.protobuf.ByteString getName() { return name_; }
-      
-      // optional int32 stores = 2;
-      public static final int STORES_FIELD_NUMBER = 2;
-      private boolean hasStores;
-      private int stores_ = 0;
-      public boolean hasStores() { return hasStores; }
-      public int getStores() { return stores_; }
-      
-      // optional int32 storefiles = 3;
-      public static final int STOREFILES_FIELD_NUMBER = 3;
-      private boolean hasStorefiles;
-      private int storefiles_ = 0;
-      public boolean hasStorefiles() { return hasStorefiles; }
-      public int getStorefiles() { return storefiles_; }
-      
-      // optional int32 storefileSizeMB = 4;
-      public static final int STOREFILESIZEMB_FIELD_NUMBER = 4;
-      private boolean hasStorefileSizeMB;
-      private int storefileSizeMB_ = 0;
-      public boolean hasStorefileSizeMB() { return hasStorefileSizeMB; }
-      public int getStorefileSizeMB() { return storefileSizeMB_; }
-      
-      // optional int32 memstoreSizeMB = 5;
-      public static final int MEMSTORESIZEMB_FIELD_NUMBER = 5;
-      private boolean hasMemstoreSizeMB;
-      private int memstoreSizeMB_ = 0;
-      public boolean hasMemstoreSizeMB() { return hasMemstoreSizeMB; }
-      public int getMemstoreSizeMB() { return memstoreSizeMB_; }
-      
-      // optional int32 storefileIndexSizeMB = 6;
-      public static final int STOREFILEINDEXSIZEMB_FIELD_NUMBER = 6;
-      private boolean hasStorefileIndexSizeMB;
-      private int storefileIndexSizeMB_ = 0;
-      public boolean hasStorefileIndexSizeMB() { return hasStorefileIndexSizeMB; }
-      public int getStorefileIndexSizeMB() { return storefileIndexSizeMB_; }
-      
-      private void initFields() {
-      }
-      public final boolean isInitialized() {
-        if (!hasName) return false;
-        return true;
-      }
-      
-      public void writeTo(com.google.protobuf.CodedOutputStream output)
-                          throws java.io.IOException {
-        getSerializedSize();
-        if (hasName()) {
-          output.writeBytes(1, getName());
-        }
-        if (hasStores()) {
-          output.writeInt32(2, getStores());
-        }
-        if (hasStorefiles()) {
-          output.writeInt32(3, getStorefiles());
-        }
-        if (hasStorefileSizeMB()) {
-          output.writeInt32(4, getStorefileSizeMB());
-        }
-        if (hasMemstoreSizeMB()) {
-          output.writeInt32(5, getMemstoreSizeMB());
-        }
-        if (hasStorefileIndexSizeMB()) {
-          output.writeInt32(6, getStorefileIndexSizeMB());
-        }
-        getUnknownFields().writeTo(output);
-      }
-      
-      private int memoizedSerializedSize = -1;
-      public int getSerializedSize() {
-        int size = memoizedSerializedSize;
-        if (size != -1) return size;
-      
-        size = 0;
-        if (hasName()) {
-          size += com.google.protobuf.CodedOutputStream
-            .computeBytesSize(1, getName());
-        }
-        if (hasStores()) {
-          size += com.google.protobuf.CodedOutputStream
-            .computeInt32Size(2, getStores());
-        }
-        if (hasStorefiles()) {
-          size += com.google.protobuf.CodedOutputStream
-            .computeInt32Size(3, getStorefiles());
-        }
-        if (hasStorefileSizeMB()) {
-          size += com.google.protobuf.CodedOutputStream
-            .computeInt32Size(4, getStorefileSizeMB());
-        }
-        if (hasMemstoreSizeMB()) {
-          size += com.google.protobuf.CodedOutputStream
-            .computeInt32Size(5, getMemstoreSizeMB());
-        }
-        if (hasStorefileIndexSizeMB()) {
-          size += com.google.protobuf.CodedOutputStream
-            .computeInt32Size(6, getStorefileIndexSizeMB());
-        }
-        size += getUnknownFields().getSerializedSize();
-        memoizedSerializedSize = size;
-        return size;
-      }
-      
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(
-          com.google.protobuf.ByteString data)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return newBuilder().mergeFrom(data).buildParsed();
-      }
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(
-          com.google.protobuf.ByteString data,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return newBuilder().mergeFrom(data, extensionRegistry)
-                 .buildParsed();
-      }
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(byte[] data)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return newBuilder().mergeFrom(data).buildParsed();
-      }
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(
-          byte[] data,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return newBuilder().mergeFrom(data, extensionRegistry)
-                 .buildParsed();
-      }
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(java.io.InputStream input)
-          throws java.io.IOException {
-        return newBuilder().mergeFrom(input).buildParsed();
-      }
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(
-          java.io.InputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        return newBuilder().mergeFrom(input, extensionRegistry)
-                 .buildParsed();
-      }
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseDelimitedFrom(java.io.InputStream input)
-          throws java.io.IOException {
-        Builder builder = newBuilder();
-        if (builder.mergeDelimitedFrom(input)) {
-          return builder.buildParsed();
-        } else {
-          return null;
-        }
-      }
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseDelimitedFrom(
-          java.io.InputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        Builder builder = newBuilder();
-        if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
-          return builder.buildParsed();
-        } else {
-          return null;
-        }
-      }
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(
-          com.google.protobuf.CodedInputStream input)
-          throws java.io.IOException {
-        return newBuilder().mergeFrom(input).buildParsed();
-      }
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        return newBuilder().mergeFrom(input, extensionRegistry)
-                 .buildParsed();
-      }
-      
-      public static Builder newBuilder() { return Builder.create(); }
-      public Builder newBuilderForType() { return newBuilder(); }
-      public static Builder newBuilder(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region prototype) {
-        return newBuilder().mergeFrom(prototype);
-      }
-      public Builder toBuilder() { return newBuilder(this); }
-      
-      public static final class Builder extends
-          com.google.protobuf.GeneratedMessage.Builder<Builder> {
-        private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region result;
-        
-        // Construct using org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.newBuilder()
-        private Builder() {}
-        
-        private static Builder create() {
-          Builder builder = new Builder();
-          builder.result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region();
-          return builder;
-        }
-        
-        protected org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region internalGetResult() {
-          return result;
-        }
-        
-        public Builder clear() {
-          if (result == null) {
-            throw new IllegalStateException(
-              "Cannot call clear() after build().");
-          }
-          result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region();
-          return this;
-        }
-        
-        public Builder clone() {
-          return create().mergeFrom(result);
-        }
-        
-        public com.google.protobuf.Descriptors.Descriptor
-            getDescriptorForType() {
-          return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.getDescriptor();
-        }
-        
-        public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region getDefaultInstanceForType() {
-          return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.getDefaultInstance();
-        }
-        
-        public boolean isInitialized() {
-          return result.isInitialized();
-        }
-        public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region build() {
-          if (result != null && !isInitialized()) {
-            throw newUninitializedMessageException(result);
-          }
-          return buildPartial();
-        }
-        
-        private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region buildParsed()
-            throws com.google.protobuf.InvalidProtocolBufferException {
-          if (!isInitialized()) {
-            throw newUninitializedMessageException(
-              result).asInvalidProtocolBufferException();
-          }
-          return buildPartial();
-        }
-        
-        public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region buildPartial() {
-          if (result == null) {
-            throw new IllegalStateException(
-              "build() has already been called on this Builder.");
-          }
-          org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region returnMe = result;
-          result = null;
-          return returnMe;
-        }
-        
-        public Builder mergeFrom(com.google.protobuf.Message other) {
-          if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region) {
-            return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region)other);
-          } else {
-            super.mergeFrom(other);
-            return this;
-          }
-        }
-        
-        public Builder mergeFrom(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region other) {
-          if (other == org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.getDefaultInstance()) return this;
-          if (other.hasName()) {
-            setName(other.getName());
-          }
-          if (other.hasStores()) {
-            setStores(other.getStores());
-          }
-          if (other.hasStorefiles()) {
-            setStorefiles(other.getStorefiles());
-          }
-          if (other.hasStorefileSizeMB()) {
-            setStorefileSizeMB(other.getStorefileSizeMB());
-          }
-          if (other.hasMemstoreSizeMB()) {
-            setMemstoreSizeMB(other.getMemstoreSizeMB());
-          }
-          if (other.hasStorefileIndexSizeMB()) {
-            setStorefileIndexSizeMB(other.getStorefileIndexSizeMB());
-          }
-          this.mergeUnknownFields(other.getUnknownFields());
-          return this;
-        }
-        
-        public Builder mergeFrom(
-            com.google.protobuf.CodedInputStream input,
-            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-            throws java.io.IOException {
-          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-            com.google.protobuf.UnknownFieldSet.newBuilder(
-              this.getUnknownFields());
-          while (true) {
-            int tag = input.readTag();
-            switch (tag) {
-              case 0:
-                this.setUnknownFields(unknownFields.build());
-                return this;
-              default: {
-                if (!parseUnknownField(input, unknownFields,
-                                       extensionRegistry, tag)) {
-                  this.setUnknownFields(unknownFields.build());
-                  return this;
-                }
-                break;
-              }
-              case 10: {
-                setName(input.readBytes());
-                break;
-              }
-              case 16: {
-                setStores(input.readInt32());
-                break;
-              }
-              case 24: {
-                setStorefiles(input.readInt32());
-                break;
-              }
-              case 32: {
-                setStorefileSizeMB(input.readInt32());
-                break;
-              }
-              case 40: {
-                setMemstoreSizeMB(input.readInt32());
-                break;
-              }
-              case 48: {
-                setStorefileIndexSizeMB(input.readInt32());
-                break;
-              }
-            }
-          }
-        }
-        
-        
-        // required bytes name = 1;
-        public boolean hasName() {
-          return result.hasName();
-        }
-        public com.google.protobuf.ByteString getName() {
-          return result.getName();
-        }
-        public Builder setName(com.google.protobuf.ByteString value) {
-          if (value == null) {
-    throw new NullPointerException();
-  }
-  result.hasName = true;
-          result.name_ = value;
-          return this;
-        }
-        public Builder clearName() {
-          result.hasName = false;
-          result.name_ = getDefaultInstance().getName();
-          return this;
-        }
-        
-        // optional int32 stores = 2;
-        public boolean hasStores() {
-          return result.hasStores();
-        }
-        public int getStores() {
-          return result.getStores();
-        }
-        public Builder setStores(int value) {
-          result.hasStores = true;
-          result.stores_ = value;
-          return this;
-        }
-        public Builder clearStores() {
-          result.hasStores = false;
-          result.stores_ = 0;
-          return this;
-        }
-        
-        // optional int32 storefiles = 3;
-        public boolean hasStorefiles() {
-          return result.hasStorefiles();
-        }
-        public int getStorefiles() {
-          return result.getStorefiles();
-        }
-        public Builder setStorefiles(int value) {
-          result.hasStorefiles = true;
-          result.storefiles_ = value;
-          return this;
-        }
-        public Builder clearStorefiles() {
-          result.hasStorefiles = false;
-          result.storefiles_ = 0;
-          return this;
-        }
-        
-        // optional int32 storefileSizeMB = 4;
-        public boolean hasStorefileSizeMB() {
-          return result.hasStorefileSizeMB();
-        }
-        public int getStorefileSizeMB() {
-          return result.getStorefileSizeMB();
-        }
-        public Builder setStorefileSizeMB(int value) {
-          result.hasStorefileSizeMB = true;
-          result.storefileSizeMB_ = value;
-          return this;
-        }
-        public Builder clearStorefileSizeMB() {
-          result.hasStorefileSizeMB = false;
-          result.storefileSizeMB_ = 0;
-          return this;
-        }
-        
-        // optional int32 memstoreSizeMB = 5;
-        public boolean hasMemstoreSizeMB() {
-          return result.hasMemstoreSizeMB();
-        }
-        public int getMemstoreSizeMB() {
-          return result.getMemstoreSizeMB();
-        }
-        public Builder setMemstoreSizeMB(int value) {
-          result.hasMemstoreSizeMB = true;
-          result.memstoreSizeMB_ = value;
-          return this;
-        }
-        public Builder clearMemstoreSizeMB() {
-          result.hasMemstoreSizeMB = false;
-          result.memstoreSizeMB_ = 0;
-          return this;
-        }
-        
-        // optional int32 storefileIndexSizeMB = 6;
-        public boolean hasStorefileIndexSizeMB() {
-          return result.hasStorefileIndexSizeMB();
-        }
-        public int getStorefileIndexSizeMB() {
-          return result.getStorefileIndexSizeMB();
-        }
-        public Builder setStorefileIndexSizeMB(int value) {
-          result.hasStorefileIndexSizeMB = true;
-          result.storefileIndexSizeMB_ = value;
-          return this;
-        }
-        public Builder clearStorefileIndexSizeMB() {
-          result.hasStorefileIndexSizeMB = false;
-          result.storefileIndexSizeMB_ = 0;
-          return this;
-        }
-        
-        // @@protoc_insertion_point(builder_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Region)
-      }
-      
-      static {
-        defaultInstance = new Region(true);
-        org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internalForceInit();
-        defaultInstance.initFields();
-      }
-      
-      // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Region)
-    }
-    
-    public static final class Node extends
-        com.google.protobuf.GeneratedMessage {
-      // Use Node.newBuilder() to construct.
-      private Node() {
-        initFields();
-      }
-      private Node(boolean noInit) {}
-      
-      private static final Node defaultInstance;
-      public static Node getDefaultInstance() {
-        return defaultInstance;
-      }
-      
-      public Node getDefaultInstanceForType() {
-        return defaultInstance;
-      }
-      
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor;
-      }
-      
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_fieldAccessorTable;
-      }
-      
-      // required string name = 1;
-      public static final int NAME_FIELD_NUMBER = 1;
-      private boolean hasName;
-      private java.lang.String name_ = "";
-      public boolean hasName() { return hasName; }
-      public java.lang.String getName() { return name_; }
-      
-      // optional int64 startCode = 2;
-      public static final int STARTCODE_FIELD_NUMBER = 2;
-      private boolean hasStartCode;
-      private long startCode_ = 0L;
-      public boolean hasStartCode() { return hasStartCode; }
-      public long getStartCode() { return startCode_; }
-      
-      // optional int32 requests = 3;
-      public static final int REQUESTS_FIELD_NUMBER = 3;
-      private boolean hasRequests;
-      private int requests_ = 0;
-      public boolean hasRequests() { return hasRequests; }
-      public int getRequests() { return requests_; }
-      
-      // optional int32 heapSizeMB = 4;
-      public static final int HEAPSIZEMB_FIELD_NUMBER = 4;
-      private boolean hasHeapSizeMB;
-      private int heapSizeMB_ = 0;
-      public boolean hasHeapSizeMB() { return hasHeapSizeMB; }
-      public int getHeapSizeMB() { return heapSizeMB_; }
-      
-      // optional int32 maxHeapSizeMB = 5;
-      public static final int MAXHEAPSIZEMB_FIELD_NUMBER = 5;
-      private boolean hasMaxHeapSizeMB;
-      private int maxHeapSizeMB_ = 0;
-      public boolean hasMaxHeapSizeMB() { return hasMaxHeapSizeMB; }
-      public int getMaxHeapSizeMB() { return maxHeapSizeMB_; }
-      
-      // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Region regions = 6;
-      public static final int REGIONS_FIELD_NUMBER = 6;
-      private java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> regions_ =
-        java.util.Collections.emptyList();
-      public java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> getRegionsList() {
-        return regions_;
-      }
-      public int getRegionsCount() { return regions_.size(); }
-      public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region getRegions(int index) {
-        return regions_.get(index);
-      }
-      
-      private void initFields() {
-      }
-      public final boolean isInitialized() {
-        if (!hasName) return false;
-        for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region element : getRegionsList()) {
-          if (!element.isInitialized()) return false;
-        }
-        return true;
-      }
-      
-      public void writeTo(com.google.protobuf.CodedOutputStream output)
-                          throws java.io.IOException {
-        getSerializedSize();
-        if (hasName()) {
-          output.writeString(1, getName());
-        }
-        if (hasStartCode()) {
-          output.writeInt64(2, getStartCode());
-        }
-        if (hasRequests()) {
-          output.writeInt32(3, getRequests());
-        }
-        if (hasHeapSizeMB()) {
-          output.writeInt32(4, getHeapSizeMB());
-        }
-        if (hasMaxHeapSizeMB()) {
-          output.writeInt32(5, getMaxHeapSizeMB());
-        }
-        for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region element : getRegionsList()) {
-          output.writeMessage(6, element);
-        }
-        getUnknownFields().writeTo(output);
-      }
-      
-      private int memoizedSerializedSize = -1;
-      public int getSerializedSize() {
-        int size = memoizedSerializedSize;
-        if (size != -1) return size;
-      
-        size = 0;
-        if (hasName()) {
-          size += com.google.protobuf.CodedOutputStream
-            .computeStringSize(1, getName());
-        }
-        if (hasStartCode()) {
-          size += com.google.protobuf.CodedOutputStream
-            .computeInt64Size(2, getStartCode());
-        }
-        if (hasRequests()) {
-          size += com.google.protobuf.CodedOutputStream
-            .computeInt32Size(3, getRequests());
-        }
-        if (hasHeapSizeMB()) {
-          size += com.google.protobuf.CodedOutputStream
-            .computeInt32Size(4, getHeapSizeMB());
-        }
-        if (hasMaxHeapSizeMB()) {
-          size += com.google.protobuf.CodedOutputStream
-            .computeInt32Size(5, getMaxHeapSizeMB());
-        }
-        for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region element : getRegionsList()) {
-          size += com.google.protobuf.CodedOutputStream
-            .computeMessageSize(6, element);
-        }
-        size += getUnknownFields().getSerializedSize();
-        memoizedSerializedSize = size;
-        return size;
-      }
-      
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(
-          com.google.protobuf.ByteString data)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return newBuilder().mergeFrom(data).buildParsed();
-      }
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(
-          com.google.protobuf.ByteString data,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return newBuilder().mergeFrom(data, extensionRegistry)
-                 .buildParsed();
-      }
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(byte[] data)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return newBuilder().mergeFrom(data).buildParsed();
-      }
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(
-          byte[] data,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return newBuilder().mergeFrom(data, extensionRegistry)
-                 .buildParsed();
-      }
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(java.io.InputStream input)
-          throws java.io.IOException {
-        return newBuilder().mergeFrom(input).buildParsed();
-      }
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(
-          java.io.InputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        return newBuilder().mergeFrom(input, extensionRegistry)
-                 .buildParsed();
-      }
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseDelimitedFrom(java.io.InputStream input)
-          throws java.io.IOException {
-        Builder builder = newBuilder();
-        if (builder.mergeDelimitedFrom(input)) {
-          return builder.buildParsed();
-        } else {
-          return null;
-        }
-      }
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseDelimitedFrom(
-          java.io.InputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        Builder builder = newBuilder();
-        if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
-          return builder.buildParsed();
-        } else {
-          return null;
-        }
-      }
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(
-          com.google.protobuf.CodedInputStream input)
-          throws java.io.IOException {
-        return newBuilder().mergeFrom(input).buildParsed();
-      }
-      public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        return newBuilder().mergeFrom(input, extensionRegistry)
-                 .buildParsed();
-      }
-      
-      public static Builder newBuilder() { return Builder.create(); }
-      public Builder newBuilderForType() { return newBuilder(); }
-      public static Builder newBuilder(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node prototype) {
-        return newBuilder().mergeFrom(prototype);
-      }
-      public Builder toBuilder() { return newBuilder(this); }
-      
-      public static final class Builder extends
-          com.google.protobuf.GeneratedMessage.Builder<Builder> {
-        private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node result;
-        
-        // Construct using org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.newBuilder()
-        private Builder() {}
-        
-        private static Builder create() {
-          Builder builder = new Builder();
-          builder.result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node();
-          return builder;
-        }
-        
-        protected org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node internalGetResult() {
-          return result;
-        }
-        
-        public Builder clear() {
-          if (result == null) {
-            throw new IllegalStateException(
-              "Cannot call clear() after build().");
-          }
-          result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node();
-          return this;
-        }
-        
-        public Builder clone() {
-          return create().mergeFrom(result);
-        }
-        
-        public com.google.protobuf.Descriptors.Descriptor
-            getDescriptorForType() {
-          return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.getDescriptor();
-        }
-        
-        public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node getDefaultInstanceForType() {
-          return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.getDefaultInstance();
-        }
-        
-        public boolean isInitialized() {
-          return result.isInitialized();
-        }
-        public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node build() {
-          if (result != null && !isInitialized()) {
-            throw newUninitializedMessageException(result);
-          }
-          return buildPartial();
-        }
-        
-        private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node buildParsed()
-            throws com.google.protobuf.InvalidProtocolBufferException {
-          if (!isInitialized()) {
-            throw newUninitializedMessageException(
-              result).asInvalidProtocolBufferException();
-          }
-          return buildPartial();
-        }
-        
-        public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node buildPartial() {
-          if (result == null) {
-            throw new IllegalStateException(
-              "build() has already been called on this Builder.");
-          }
-          if (result.regions_ != java.util.Collections.EMPTY_LIST) {
-            result.regions_ =
-              java.util.Collections.unmodifiableList(result.regions_);
-          }
-          org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node returnMe = result;
-          result = null;
-          return returnMe;
-        }
-        
-        public Builder mergeFrom(com.google.protobuf.Message other) {
-          if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node) {
-            return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node)other);
-          } else {
-            super.mergeFrom(other);
-            return this;
-          }
-        }
-        
-        public Builder mergeFrom(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node other) {
-          if (other == org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.getDefaultInstance()) return this;
-          if (other.hasName()) {
-            setName(other.getName());
-          }
-          if (other.hasStartCode()) {
-            setStartCode(other.getStartCode());
-          }
-          if (other.hasRequests()) {
-            setRequests(other.getRequests());
-          }
-          if (other.hasHeapSizeMB()) {
-            setHeapSizeMB(other.getHeapSizeMB());
-          }
-          if (other.hasMaxHeapSizeMB()) {
-            setMaxHeapSizeMB(other.getMaxHeapSizeMB());
-          }
-          if (!other.regions_.isEmpty()) {
-            if (result.regions_.isEmpty()) {
-              result.regions_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region>();
-            }
-            result.regions_.addAll(other.regions_);
-          }
-          this.mergeUnknownFields(other.getUnknownFields());
-          return this;
-        }
-        
-        public Builder mergeFrom(
-            com.google.protobuf.CodedInputStream input,
-            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-            throws java.io.IOException {
-          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-            com.google.protobuf.UnknownFieldSet.newBuilder(
-              this.getUnknownFields());
-          while (true) {
-            int tag = input.readTag();
-            switch (tag) {
-              case 0:
-                this.setUnknownFields(unknownFields.build());
-                return this;
-              default: {
-                if (!parseUnknownField(input, unknownFields,
-                                       extensionRegistry, tag)) {
-                  this.setUnknownFields(unknownFields.build());
-                  return this;
-                }
-                break;
-              }
-              case 10: {
-                setName(input.readString());
-                break;
-              }
-              case 16: {
-                setStartCode(input.readInt64());
-                break;
-              }
-              case 24: {
-                setRequests(input.readInt32());
-                break;
-              }
-              case 32: {
-                setHeapSizeMB(input.readInt32());
-                break;
-              }
-              case 40: {
-                setMaxHeapSizeMB(input.readInt32());
-                break;
-              }
-              case 50: {
-                org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder subBuilder = org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.newBuilder();
-                input.readMessage(subBuilder, extensionRegistry);
-                addRegions(subBuilder.buildPartial());
-                break;
-              }
-            }
-          }
-        }
-        
-        
-        // required string name = 1;
-        public boolean hasName() {
-          return result.hasName();
-        }
-        public java.lang.String getName() {
-          return result.getName();
-        }
-        public Builder setName(java.lang.String value) {
-          if (value == null) {
-    throw new NullPointerException();
-  }
-  result.hasName = true;
-          result.name_ = value;
-          return this;
-        }
-        public Builder clearName() {
-          result.hasName = false;
-          result.name_ = getDefaultInstance().getName();
-          return this;
-        }
-        
-        // optional int64 startCode = 2;
-        public boolean hasStartCode() {
-          return result.hasStartCode();
-        }
-        public long getStartCode() {
-          return result.getStartCode();
-        }
-        public Builder setStartCode(long value) {
-          result.hasStartCode = true;
-          result.startCode_ = value;
-          return this;
-        }
-        public Builder clearStartCode() {
-          result.hasStartCode = false;
-          result.startCode_ = 0L;
-          return this;
-        }
-        
-        // optional int32 requests = 3;
-        public boolean hasRequests() {
-          return result.hasRequests();
-        }
-        public int getRequests() {
-          return result.getRequests();
-        }
-        public Builder setRequests(int value) {
-          result.hasRequests = true;
-          result.requests_ = value;
-          return this;
-        }
-        public Builder clearRequests() {
-          result.hasRequests = false;
-          result.requests_ = 0;
-          return this;
-        }
-        
-        // optional int32 heapSizeMB = 4;
-        public boolean hasHeapSizeMB() {
-          return result.hasHeapSizeMB();
-        }
-        public int getHeapSizeMB() {
-          return result.getHeapSizeMB();
-        }
-        public Builder setHeapSizeMB(int value) {
-          result.hasHeapSizeMB = true;
-          result.heapSizeMB_ = value;
-          return this;
-        }
-        public Builder clearHeapSizeMB() {
-          result.hasHeapSizeMB = false;
-          result.heapSizeMB_ = 0;
-          return this;
-        }
-        
-        // optional int32 maxHeapSizeMB = 5;
-        public boolean hasMaxHeapSizeMB() {
-          return result.hasMaxHeapSizeMB();
-        }
-        public int getMaxHeapSizeMB() {
-          return result.getMaxHeapSizeMB();
-        }
-        public Builder setMaxHeapSizeMB(int value) {
-          result.hasMaxHeapSizeMB = true;
-          result.maxHeapSizeMB_ = value;
-          return this;
-        }
-        public Builder clearMaxHeapSizeMB() {
-          result.hasMaxHeapSizeMB = false;
-          result.maxHeapSizeMB_ = 0;
-          return this;
-        }
-        
-        // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Region regions = 6;
-        public java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> getRegionsList() {
-          return java.util.Collections.unmodifiableList(result.regions_);
-        }
-        public int getRegionsCount() {
-          return result.getRegionsCount();
-        }
-        public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region getRegions(int index) {
-          return result.getRegions(index);
-        }
-        public Builder setRegions(int index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region value) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          result.regions_.set(index, value);
-          return this;
-        }
-        public Builder setRegions(int index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder builderForValue) {
-          result.regions_.set(index, builderForValue.build());
-          return this;
-        }
-        public Builder addRegions(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region value) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          if (result.regions_.isEmpty()) {
-            result.regions_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region>();
-          }
-          result.regions_.add(value);
-          return this;
-        }
-        public Builder addRegions(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder builderForValue) {
-          if (result.regions_.isEmpty()) {
-            result.regions_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region>();
-          }
-          result.regions_.add(builderForValue.build());
-          return this;
-        }
-        public Builder addAllRegions(
-            java.lang.Iterable<? extends org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> values) {
-          if (result.regions_.isEmpty()) {
-            result.regions_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region>();
-          }
-          super.addAll(values, result.regions_);
-          return this;
-        }
-        public Builder clearRegions() {
-          result.regions_ = java.util.Collections.emptyList();
-          return this;
-        }
-        
-        // @@protoc_insertion_point(builder_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node)
-      }
-      
-      static {
-        defaultInstance = new Node(true);
-        org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internalForceInit();
-        defaultInstance.initFields();
-      }
-      
-      // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node)
-    }
-    
-    // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node liveNodes = 1;
-    public static final int LIVENODES_FIELD_NUMBER = 1;
-    private java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node> liveNodes_ =
-      java.util.Collections.emptyList();
-    public java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node> getLiveNodesList() {
-      return liveNodes_;
-    }
-    public int getLiveNodesCount() { return liveNodes_.size(); }
-    public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node getLiveNodes(int index) {
-      return liveNodes_.get(index);
-    }
-    
-    // repeated string deadNodes = 2;
-    public static final int DEADNODES_FIELD_NUMBER = 2;
-    private java.util.List<java.lang.String> deadNodes_ =
-      java.util.Collections.emptyList();
-    public java.util.List<java.lang.String> getDeadNodesList() {
-      return deadNodes_;
-    }
-    public int getDeadNodesCount() { return deadNodes_.size(); }
-    public java.lang.String getDeadNodes(int index) {
-      return deadNodes_.get(index);
-    }
-    
-    // optional int32 regions = 3;
-    public static final int REGIONS_FIELD_NUMBER = 3;
-    private boolean hasRegions;
-    private int regions_ = 0;
-    public boolean hasRegions() { return hasRegions; }
-    public int getRegions() { return regions_; }
-    
-    // optional int32 requests = 4;
-    public static final int REQUESTS_FIELD_NUMBER = 4;
-    private boolean hasRequests;
-    private int requests_ = 0;
-    public boolean hasRequests() { return hasRequests; }
-    public int getRequests() { return requests_; }
-    
-    // optional double averageLoad = 5;
-    public static final int AVERAGELOAD_FIELD_NUMBER = 5;
-    private boolean hasAverageLoad;
-    private double averageLoad_ = 0D;
-    public boolean hasAverageLoad() { return hasAverageLoad; }
-    public double getAverageLoad() { return averageLoad_; }
-    
-    private void initFields() {
-    }
-    public final boolean isInitialized() {
-      for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node element : getLiveNodesList()) {
-        if (!element.isInitialized()) return false;
-      }
-      return true;
-    }
-    
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node element : getLiveNodesList()) {
-        output.writeMessage(1, element);
-      }
-      for (java.lang.String element : getDeadNodesList()) {
-        output.writeString(2, element);
-      }
-      if (hasRegions()) {
-        output.writeInt32(3, getRegions());
-      }
-      if (hasRequests()) {
-        output.writeInt32(4, getRequests());
-      }
-      if (hasAverageLoad()) {
-        output.writeDouble(5, getAverageLoad());
-      }
-      getUnknownFields().writeTo(output);
-    }
-    
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-    
-      size = 0;
-      for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node element : getLiveNodesList()) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeMessageSize(1, element);
-      }
-      {
-        int dataSize = 0;
-        for (java.lang.String element : getDeadNodesList()) {
-          dataSize += com.google.protobuf.CodedOutputStream
-            .computeStringSizeNoTag(element);
-        }
-        size += dataSize;
-        size += 1 * getDeadNodesList().size();
-      }
-      if (hasRegions()) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeInt32Size(3, getRegions());
-      }
-      if (hasRequests()) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeInt32Size(4, getRequests());
-      }
-      if (hasAverageLoad()) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeDoubleSize(5, getAverageLoad());
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-    
-    public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data).buildParsed();
-    }
-    public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data, extensionRegistry)
-               .buildParsed();
-    }
-    public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data).buildParsed();
-    }
-    public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data, extensionRegistry)
-               .buildParsed();
-    }
-    public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return newBuilder().mergeFrom(input).buildParsed();
-    }
-    public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return newBuilder().mergeFrom(input, extensionRegistry)
-               .buildParsed();
-    }
-    public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      Builder builder = newBuilder();
-      if (builder.mergeDelimitedFrom(input)) {
-        return builder.buildParsed();
-      } else {
-        return null;
-      }
-    }
-    public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      Builder builder = newBuilder();
-      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
-        return builder.buildParsed();
-      } else {
-        return null;
-      }
-    }
-    public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return newBuilder().mergeFrom(input).buildParsed();
-    }
-    public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return newBuilder().mergeFrom(input, extensionRegistry)
-               .buildParsed();
-    }
-    
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-    
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder> {
-      private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus result;
-      
-      // Construct using org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.newBuilder()
-      private Builder() {}
-      
-      private static Builder create() {
-        Builder builder = new Builder();
-        builder.result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus();
-        return builder;
-      }
-      
-      protected org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus internalGetResult() {
-        return result;
-      }
-      
-      public Builder clear() {
-        if (result == null) {
-          throw new IllegalStateException(
-            "Cannot call clear() after build().");
-        }
-        result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus();
-        return this;
-      }
-      
-      public Builder clone() {
-        return create().mergeFrom(result);
-      }
-      
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.getDescriptor();
-      }
-      
-      public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.getDefaultInstance();
-      }
-      
-      public boolean isInitialized() {
-        return result.isInitialized();
-      }
-      public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus build() {
-        if (result != null && !isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return buildPartial();
-      }
-      
-      private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus buildParsed()
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        if (!isInitialized()) {
-          throw newUninitializedMessageException(
-            result).asInvalidProtocolBufferException();
-        }
-        return buildPartial();
-      }
-      
-      public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus buildPartial() {
-        if (result == null) {
-          throw new IllegalStateException(
-            "build() has already been called on this Builder.");
-        }
-        if (result.liveNodes_ != java.util.Collections.EMPTY_LIST) {
-          result.liveNodes_ =
-            java.util.Collections.unmodifiableList(result.liveNodes_);
-        }
-        if (result.deadNodes_ != java.util.Collections.EMPTY_LIST) {
-          result.deadNodes_ =
-            java.util.Collections.unmodifiableList(result.deadNodes_);
-        }
-        org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus returnMe = result;
-        result = null;
-        return returnMe;
-      }
-      
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus) {
-          return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-      
-      public Builder mergeFrom(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus other) {
-        if (other == org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.getDefaultInstance()) return this;
-        if (!other.liveNodes_.isEmpty()) {
-          if (result.liveNodes_.isEmpty()) {
-            result.liveNodes_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node>();
-          }
-          result.liveNodes_.addAll(other.liveNodes_);
-        }
-        if (!other.deadNodes_.isEmpty()) {
-          if (result.deadNodes_.isEmpty()) {
-            result.deadNodes_ = new java.util.ArrayList<java.lang.String>();
-          }
-          result.deadNodes_.addAll(other.deadNodes_);
-        }
-        if (other.hasRegions()) {
-          setRegions(other.getRegions());
-        }
-        if (other.hasRequests()) {
-          setRequests(other.getRequests());
-        }
-        if (other.hasAverageLoad()) {
-          setAverageLoad(other.getAverageLoad());
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-      
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder(
-            this.getUnknownFields());
-        while (true) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              this.setUnknownFields(unknownFields.build());
-              return this;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                this.setUnknownFields(unknownFields.build());
-                return this;
-              }
-              break;
-            }
-            case 10: {
-              org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.Builder subBuilder = org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.newBuilder();
-              input.readMessage(subBuilder, extensionRegistry);
-              addLiveNodes(subBuilder.buildPartial());
-              break;
-            }
-            case 18: {
-              addDeadNodes(input.readString());
-              break;
-            }
-            case 24: {
-              setRegions(input.readInt32());
-              break;
-            }
-            case 32: {
-              setRequests(input.readInt32());
-              break;
-            }
-            case 41: {
-              setAverageLoad(input.readDouble());
-              break;
-            }
-          }
-        }
-      }
-      
-      
-      // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node liveNodes = 1;
-      public java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node> getLiveNodesList() {
-        return java.util.Collections.unmodifiableList(result.liveNodes_);
-      }
-      public int getLiveNodesCount() {
-        return result.getLiveNodesCount();
-      }
-      public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node getLiveNodes(int index) {
-        return result.getLiveNodes(index);
-      }
-      public Builder setLiveNodes(int index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node value) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        result.liveNodes_.set(index, value);
-        return this;
-      }
-      public Builder setLiveNodes(int index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.Builder builderForValue) {
-        result.liveNodes_.set(index, builderForValue.build());
-        return this;
-      }
-      public Builder addLiveNodes(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node value) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        if (result.liveNodes_.isEmpty()) {
-          result.liveNodes_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node>();
-        }
-        result.liveNodes_.add(value);
-        return this;
-      }
-      public Builder addLiveNodes(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.Builder builderForValue) {
-        if (result.liveNodes_.isEmpty()) {
-          result.liveNodes_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node>();
-        }
-        result.liveNodes_.add(builderForValue.build());
-        return this;
-      }
-      public Builder addAllLiveNodes(
-          java.lang.Iterable<? extends org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node> values) {
-        if (result.liveNodes_.isEmpty()) {
-          result.liveNodes_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node>();
-        }
-        super.addAll(values, result.liveNodes_);
-        return this;
-      }
-      public Builder clearLiveNodes() {
-        result.liveNodes_ = java.util.Collections.emptyList();
-        return this;
-      }
-      
-      // repeated string deadNodes = 2;
-      public java.util.List<java.lang.String> getDeadNodesList() {
-        return java.util.Collections.unmodifiableList(result.deadNodes_);
-      }
-      public int getDeadNodesCount() {
-        return result.getDeadNodesCount();
-      }
-      public java.lang.String getDeadNodes(int index) {
-        return result.getDeadNodes(index);
-      }
-      public Builder setDeadNodes(int index, java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  result.deadNodes_.set(index, value);
-        return this;
-      }
-      public Builder addDeadNodes(java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  if (result.deadNodes_.isEmpty()) {
-          result.deadNodes_ = new java.util.ArrayList<java.lang.String>();
-        }
-        result.deadNodes_.add(value);
-        return this;
-      }
-      public Builder addAllDeadNodes(
-          java.lang.Iterable<? extends java.lang.String> values) {
-        if (result.deadNodes_.isEmpty()) {
-          result.deadNodes_ = new java.util.ArrayList<java.lang.String>();
-        }
-        super.addAll(values, result.deadNodes_);
-        return this;
-      }
-      public Builder clearDeadNodes() {
-        result.deadNodes_ = java.util.Collections.emptyList();
-        return this;
-      }
-      
-      // optional int32 regions = 3;
-      public boolean hasRegions() {
-        return result.hasRegions();
-      }
-      public int getRegions() {
-        return result.getRegions();
-      }
-      public Builder setRegions(int value) {
-        result.hasRegions = true;
-        result.regions_ = value;
-        return this;
-      }
-      public Builder clearRegions() {
-        result.hasRegions = false;
-        result.regions_ = 0;
-        return this;
-      }
-      
-      // optional int32 requests = 4;
-      public boolean hasRequests() {
-        return result.hasRequests();
-      }
-      public int getRequests() {
-        return result.getRequests();
-      }
-      public Builder setRequests(int value) {
-        result.hasRequests = true;
-        result.requests_ = value;
-        return this;
-      }
-      public Builder clearRequests() {
-        result.hasRequests = false;
-        result.requests_ = 0;
-        return this;
-      }
-      
-      // optional double averageLoad = 5;
-      public boolean hasAverageLoad() {
-        return result.hasAverageLoad();
-      }
-      public double getAverageLoad() {
-        return result.getAverageLoad();
-      }
-      public Builder setAverageLoad(double value) {
-        result.hasAverageLoad = true;
-        result.averageLoad_ = value;
-        return this;
-      }
-      public Builder clearAverageLoad() {
-        result.hasAverageLoad = false;
-        result.averageLoad_ = 0D;
-        return this;
-      }
-      
-      // @@protoc_insertion_point(builder_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus)
-    }
-    
-    static {
-      defaultInstance = new StorageClusterStatus(true);
-      org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internalForceInit();
-      defaultInstance.initFields();
-    }
-    
-    // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus)
-  }
-  
-  private static com.google.protobuf.Descriptors.Descriptor
-    internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
-      internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_fieldAccessorTable;
-  private static com.google.protobuf.Descriptors.Descriptor
-    internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
-      internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_fieldAccessorTable;
-  private static com.google.protobuf.Descriptors.Descriptor
-    internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
-      internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_fieldAccessorTable;
-  
-  public static com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n!StorageClusterStatusMessage.proto\022/org" +
-      ".apache.hadoop.hbase.rest.protobuf.gener" +
-      "ated\"\222\004\n\024StorageClusterStatus\022]\n\tliveNod" +
-      "es\030\001 \003(\0132J.org.apache.hadoop.hbase.rest." +
-      "protobuf.generated.StorageClusterStatus." +
-      "Node\022\021\n\tdeadNodes\030\002 \003(\t\022\017\n\007regions\030\003 \001(\005" +
-      "\022\020\n\010requests\030\004 \001(\005\022\023\n\013averageLoad\030\005 \001(\001\032" +
-      "\211\001\n\006Region\022\014\n\004name\030\001 \002(\014\022\016\n\006stores\030\002 \001(\005" +
-      "\022\022\n\nstorefiles\030\003 \001(\005\022\027\n\017storefileSizeMB\030" +
-      "\004 \001(\005\022\026\n\016memstoreSizeMB\030\005 \001(\005\022\034\n\024storefi",
-      "leIndexSizeMB\030\006 \001(\005\032\303\001\n\004Node\022\014\n\004name\030\001 \002" +
-      "(\t\022\021\n\tstartCode\030\002 \001(\003\022\020\n\010requests\030\003 \001(\005\022" +
-      "\022\n\nheapSizeMB\030\004 \001(\005\022\025\n\rmaxHeapSizeMB\030\005 \001" +
-      "(\005\022]\n\007regions\030\006 \003(\0132L.org.apache.hadoop." +
-      "hbase.rest.protobuf.generated.StorageClu" +
-      "sterStatus.Region"
-    };
-    com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-      new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
-        public com.google.protobuf.ExtensionRegistry assignDescriptors(
-            com.google.protobuf.Descriptors.FileDescriptor root) {
-          descriptor = root;
-          internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor =
-            getDescriptor().getMessageTypes().get(0);
-          internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor,
-              new java.lang.String[] { "LiveNodes", "DeadNodes", "Regions", "Requests", "AverageLoad", },
-              org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.class,
-              org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Builder.class);
-          internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor =
-            internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor.getNestedTypes().get(0);
-          internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor,
-              new java.lang.String[] { "Name", "Stores", "Storefiles", "StorefileSizeMB", "MemstoreSizeMB", "StorefileIndexSizeMB", },
-              org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.class,
-              org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder.class);
-          internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor =
-            internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor.getNestedTypes().get(1);
-          internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor,
-              new java.lang.String[] { "Name", "StartCode", "Requests", "HeapSizeMB", "MaxHeapSizeMB", "Regions", },
-              org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.class,
-              org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.Builder.class);
-          return null;
-        }
-      };
-    com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new com.google.protobuf.Descriptors.FileDescriptor[] {
-        }, assigner);
-  }
-  
-  public static void internalForceInit() {}
-  
-  // @@protoc_insertion_point(outer_class_scope)
-}



Mime
View raw message