Return-Path: X-Original-To: apmail-hbase-commits-archive@www.apache.org Delivered-To: apmail-hbase-commits-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 1810911689 for ; Wed, 14 May 2014 00:28:10 +0000 (UTC) Received: (qmail 42393 invoked by uid 500); 14 May 2014 00:28:09 -0000 Delivered-To: apmail-hbase-commits-archive@hbase.apache.org Received: (qmail 42346 invoked by uid 500); 14 May 2014 00:28:09 -0000 Mailing-List: contact commits-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@hbase.apache.org Delivered-To: mailing list commits@hbase.apache.org Received: (qmail 42339 invoked by uid 99); 14 May 2014 00:28:09 -0000 Received: from Unknown (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 14 May 2014 00:28:09 +0000 X-ASF-Spam-Status: No, hits=-2000.0 required=5.0 tests=ALL_TRUSTED X-Spam-Check-By: apache.org Received: from [140.211.11.4] (HELO eris.apache.org) (140.211.11.4) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 14 May 2014 00:28:05 +0000 Received: from eris.apache.org (localhost [127.0.0.1]) by eris.apache.org (Postfix) with ESMTP id 1A23F2388BA6; Wed, 14 May 2014 00:27:21 +0000 (UTC) Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit Subject: svn commit: r1594423 [9/17] - in /hbase/branches/0.89-fb: ./ bin/ src/main/java/org/apache/hadoop/hbase/rest/ src/main/java/org/apache/hadoop/hbase/rest/client/ src/main/java/org/apache/hadoop/hbase/rest/metrics/ src/main/java/org/apache/hadoop/hbase/r... Date: Wed, 14 May 2014 00:27:04 -0000 To: commits@hbase.apache.org From: liyin@apache.org X-Mailer: svnmailer-1.0.9 Message-Id: <20140514002721.1A23F2388BA6@eris.apache.org> X-Virus-Checked: Checked by ClamAV on apache.org Modified: hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/TableSchemaMessage.java URL: http://svn.apache.org/viewvc/hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/TableSchemaMessage.java?rev=1594423&r1=1594422&r2=1594423&view=diff ============================================================================== --- hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/TableSchemaMessage.java (original) +++ hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/TableSchemaMessage.java Wed May 14 00:26:57 2014 @@ -1,949 +0,0 @@ -// Generated by the protocol buffer compiler. DO NOT EDIT! -// source: TableSchemaMessage.proto - -package org.apache.hadoop.hbase.rest.protobuf.generated; - -public final class TableSchemaMessage { - private TableSchemaMessage() {} - public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { - } - public static final class TableSchema extends - com.google.protobuf.GeneratedMessage { - // Use TableSchema.newBuilder() to construct. - private TableSchema() { - initFields(); - } - private TableSchema(boolean noInit) {} - - private static final TableSchema defaultInstance; - public static TableSchema getDefaultInstance() { - return defaultInstance; - } - - public TableSchema getDefaultInstanceForType() { - return defaultInstance; - } - - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_fieldAccessorTable; - } - - public static final class Attribute extends - com.google.protobuf.GeneratedMessage { - // Use Attribute.newBuilder() to construct. - private Attribute() { - initFields(); - } - private Attribute(boolean noInit) {} - - private static final Attribute defaultInstance; - public static Attribute getDefaultInstance() { - return defaultInstance; - } - - public Attribute getDefaultInstanceForType() { - return defaultInstance; - } - - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_Attribute_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_Attribute_fieldAccessorTable; - } - - // required string name = 1; - public static final int NAME_FIELD_NUMBER = 1; - private boolean hasName; - private java.lang.String name_ = ""; - public boolean hasName() { return hasName; } - public java.lang.String getName() { return name_; } - - // required string value = 2; - public static final int VALUE_FIELD_NUMBER = 2; - private boolean hasValue; - private java.lang.String value_ = ""; - public boolean hasValue() { return hasValue; } - public java.lang.String getValue() { return value_; } - - private void initFields() { - } - public final boolean isInitialized() { - if (!hasName) return false; - if (!hasValue) return false; - return true; - } - - public void writeTo(com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - getSerializedSize(); - if (hasName()) { - output.writeString(1, getName()); - } - if (hasValue()) { - output.writeString(2, getValue()); - } - getUnknownFields().writeTo(output); - } - - private int memoizedSerializedSize = -1; - public int getSerializedSize() { - int size = memoizedSerializedSize; - if (size != -1) return size; - - size = 0; - if (hasName()) { - size += com.google.protobuf.CodedOutputStream - .computeStringSize(1, getName()); - } - if (hasValue()) { - size += com.google.protobuf.CodedOutputStream - .computeStringSize(2, getValue()); - } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; - return size; - } - - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data).buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data).buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseFrom( - byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseFrom(java.io.InputStream input) - throws java.io.IOException { - return newBuilder().mergeFrom(input).buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return newBuilder().mergeFrom(input, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - Builder builder = newBuilder(); - if (builder.mergeDelimitedFrom(input)) { - return builder.buildParsed(); - } else { - return null; - } - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseDelimitedFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - Builder builder = newBuilder(); - if (builder.mergeDelimitedFrom(input, extensionRegistry)) { - return builder.buildParsed(); - } else { - return null; - } - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseFrom( - com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return newBuilder().mergeFrom(input).buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return newBuilder().mergeFrom(input, extensionRegistry) - .buildParsed(); - } - - public static Builder newBuilder() { return Builder.create(); } - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute prototype) { - return newBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { return newBuilder(this); } - - public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder { - private org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute result; - - // Construct using org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute.newBuilder() - private Builder() {} - - private static Builder create() { - Builder builder = new Builder(); - builder.result = new org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute(); - return builder; - } - - protected org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute internalGetResult() { - return result; - } - - public Builder clear() { - if (result == null) { - throw new IllegalStateException( - "Cannot call clear() after build()."); - } - result = new org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute(); - return this; - } - - public Builder clone() { - return create().mergeFrom(result); - } - - public com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute.getDescriptor(); - } - - public org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute getDefaultInstanceForType() { - return org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute.getDefaultInstance(); - } - - public boolean isInitialized() { - return result.isInitialized(); - } - public org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute build() { - if (result != null && !isInitialized()) { - throw newUninitializedMessageException(result); - } - return buildPartial(); - } - - private org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute buildParsed() - throws com.google.protobuf.InvalidProtocolBufferException { - if (!isInitialized()) { - throw newUninitializedMessageException( - result).asInvalidProtocolBufferException(); - } - return buildPartial(); - } - - public org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute buildPartial() { - if (result == null) { - throw new IllegalStateException( - "build() has already been called on this Builder."); - } - org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute returnMe = result; - result = null; - return returnMe; - } - - public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute) { - return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute other) { - if (other == org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute.getDefaultInstance()) return this; - if (other.hasName()) { - setName(other.getName()); - } - if (other.hasValue()) { - setValue(other.getValue()); - } - this.mergeUnknownFields(other.getUnknownFields()); - return this; - } - - public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder( - this.getUnknownFields()); - while (true) { - int tag = input.readTag(); - switch (tag) { - case 0: - this.setUnknownFields(unknownFields.build()); - return this; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - this.setUnknownFields(unknownFields.build()); - return this; - } - break; - } - case 10: { - setName(input.readString()); - break; - } - case 18: { - setValue(input.readString()); - break; - } - } - } - } - - - // required string name = 1; - public boolean hasName() { - return result.hasName(); - } - public java.lang.String getName() { - return result.getName(); - } - public Builder setName(java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - result.hasName = true; - result.name_ = value; - return this; - } - public Builder clearName() { - result.hasName = false; - result.name_ = getDefaultInstance().getName(); - return this; - } - - // required string value = 2; - public boolean hasValue() { - return result.hasValue(); - } - public java.lang.String getValue() { - return result.getValue(); - } - public Builder setValue(java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - result.hasValue = true; - result.value_ = value; - return this; - } - public Builder clearValue() { - result.hasValue = false; - result.value_ = getDefaultInstance().getValue(); - return this; - } - - // @@protoc_insertion_point(builder_scope:org.apache.hadoop.hbase.rest.protobuf.generated.TableSchema.Attribute) - } - - static { - defaultInstance = new Attribute(true); - org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.internalForceInit(); - defaultInstance.initFields(); - } - - // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.TableSchema.Attribute) - } - - // optional string name = 1; - public static final int NAME_FIELD_NUMBER = 1; - private boolean hasName; - private java.lang.String name_ = ""; - public boolean hasName() { return hasName; } - public java.lang.String getName() { return name_; } - - // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.TableSchema.Attribute attrs = 2; - public static final int ATTRS_FIELD_NUMBER = 2; - private java.util.List attrs_ = - java.util.Collections.emptyList(); - public java.util.List getAttrsList() { - return attrs_; - } - public int getAttrsCount() { return attrs_.size(); } - public org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute getAttrs(int index) { - return attrs_.get(index); - } - - // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchema columns = 3; - public static final int COLUMNS_FIELD_NUMBER = 3; - private java.util.List columns_ = - java.util.Collections.emptyList(); - public java.util.List getColumnsList() { - return columns_; - } - public int getColumnsCount() { return columns_.size(); } - public org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema getColumns(int index) { - return columns_.get(index); - } - - // optional bool inMemory = 4; - public static final int INMEMORY_FIELD_NUMBER = 4; - private boolean hasInMemory; - private boolean inMemory_ = false; - public boolean hasInMemory() { return hasInMemory; } - public boolean getInMemory() { return inMemory_; } - - // optional bool readOnly = 5; - public static final int READONLY_FIELD_NUMBER = 5; - private boolean hasReadOnly; - private boolean readOnly_ = false; - public boolean hasReadOnly() { return hasReadOnly; } - public boolean getReadOnly() { return readOnly_; } - - private void initFields() { - } - public final boolean isInitialized() { - for (org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute element : getAttrsList()) { - if (!element.isInitialized()) return false; - } - for (org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema element : getColumnsList()) { - if (!element.isInitialized()) return false; - } - return true; - } - - public void writeTo(com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - getSerializedSize(); - if (hasName()) { - output.writeString(1, getName()); - } - for (org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute element : getAttrsList()) { - output.writeMessage(2, element); - } - for (org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema element : getColumnsList()) { - output.writeMessage(3, element); - } - if (hasInMemory()) { - output.writeBool(4, getInMemory()); - } - if (hasReadOnly()) { - output.writeBool(5, getReadOnly()); - } - getUnknownFields().writeTo(output); - } - - private int memoizedSerializedSize = -1; - public int getSerializedSize() { - int size = memoizedSerializedSize; - if (size != -1) return size; - - size = 0; - if (hasName()) { - size += com.google.protobuf.CodedOutputStream - .computeStringSize(1, getName()); - } - for (org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute element : getAttrsList()) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(2, element); - } - for (org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema element : getColumnsList()) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(3, element); - } - if (hasInMemory()) { - size += com.google.protobuf.CodedOutputStream - .computeBoolSize(4, getInMemory()); - } - if (hasReadOnly()) { - size += com.google.protobuf.CodedOutputStream - .computeBoolSize(5, getReadOnly()); - } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; - return size; - } - - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data).buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data).buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseFrom( - byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseFrom(java.io.InputStream input) - throws java.io.IOException { - return newBuilder().mergeFrom(input).buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return newBuilder().mergeFrom(input, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - Builder builder = newBuilder(); - if (builder.mergeDelimitedFrom(input)) { - return builder.buildParsed(); - } else { - return null; - } - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseDelimitedFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - Builder builder = newBuilder(); - if (builder.mergeDelimitedFrom(input, extensionRegistry)) { - return builder.buildParsed(); - } else { - return null; - } - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseFrom( - com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return newBuilder().mergeFrom(input).buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return newBuilder().mergeFrom(input, extensionRegistry) - .buildParsed(); - } - - public static Builder newBuilder() { return Builder.create(); } - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema prototype) { - return newBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { return newBuilder(this); } - - public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder { - private org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema result; - - // Construct using org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.newBuilder() - private Builder() {} - - private static Builder create() { - Builder builder = new Builder(); - builder.result = new org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema(); - return builder; - } - - protected org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema internalGetResult() { - return result; - } - - public Builder clear() { - if (result == null) { - throw new IllegalStateException( - "Cannot call clear() after build()."); - } - result = new org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema(); - return this; - } - - public Builder clone() { - return create().mergeFrom(result); - } - - public com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.getDescriptor(); - } - - public org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema getDefaultInstanceForType() { - return org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.getDefaultInstance(); - } - - public boolean isInitialized() { - return result.isInitialized(); - } - public org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema build() { - if (result != null && !isInitialized()) { - throw newUninitializedMessageException(result); - } - return buildPartial(); - } - - private org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema buildParsed() - throws com.google.protobuf.InvalidProtocolBufferException { - if (!isInitialized()) { - throw newUninitializedMessageException( - result).asInvalidProtocolBufferException(); - } - return buildPartial(); - } - - public org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema buildPartial() { - if (result == null) { - throw new IllegalStateException( - "build() has already been called on this Builder."); - } - if (result.attrs_ != java.util.Collections.EMPTY_LIST) { - result.attrs_ = - java.util.Collections.unmodifiableList(result.attrs_); - } - if (result.columns_ != java.util.Collections.EMPTY_LIST) { - result.columns_ = - java.util.Collections.unmodifiableList(result.columns_); - } - org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema returnMe = result; - result = null; - return returnMe; - } - - public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema) { - return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema other) { - if (other == org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.getDefaultInstance()) return this; - if (other.hasName()) { - setName(other.getName()); - } - if (!other.attrs_.isEmpty()) { - if (result.attrs_.isEmpty()) { - result.attrs_ = new java.util.ArrayList(); - } - result.attrs_.addAll(other.attrs_); - } - if (!other.columns_.isEmpty()) { - if (result.columns_.isEmpty()) { - result.columns_ = new java.util.ArrayList(); - } - result.columns_.addAll(other.columns_); - } - if (other.hasInMemory()) { - setInMemory(other.getInMemory()); - } - if (other.hasReadOnly()) { - setReadOnly(other.getReadOnly()); - } - this.mergeUnknownFields(other.getUnknownFields()); - return this; - } - - public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder( - this.getUnknownFields()); - while (true) { - int tag = input.readTag(); - switch (tag) { - case 0: - this.setUnknownFields(unknownFields.build()); - return this; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - this.setUnknownFields(unknownFields.build()); - return this; - } - break; - } - case 10: { - setName(input.readString()); - break; - } - case 18: { - org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute.Builder subBuilder = org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute.newBuilder(); - input.readMessage(subBuilder, extensionRegistry); - addAttrs(subBuilder.buildPartial()); - break; - } - case 26: { - org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Builder subBuilder = org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.newBuilder(); - input.readMessage(subBuilder, extensionRegistry); - addColumns(subBuilder.buildPartial()); - break; - } - case 32: { - setInMemory(input.readBool()); - break; - } - case 40: { - setReadOnly(input.readBool()); - break; - } - } - } - } - - - // optional string name = 1; - public boolean hasName() { - return result.hasName(); - } - public java.lang.String getName() { - return result.getName(); - } - public Builder setName(java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - result.hasName = true; - result.name_ = value; - return this; - } - public Builder clearName() { - result.hasName = false; - result.name_ = getDefaultInstance().getName(); - return this; - } - - // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.TableSchema.Attribute attrs = 2; - public java.util.List getAttrsList() { - return java.util.Collections.unmodifiableList(result.attrs_); - } - public int getAttrsCount() { - return result.getAttrsCount(); - } - public org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute getAttrs(int index) { - return result.getAttrs(index); - } - public Builder setAttrs(int index, org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute value) { - if (value == null) { - throw new NullPointerException(); - } - result.attrs_.set(index, value); - return this; - } - public Builder setAttrs(int index, org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute.Builder builderForValue) { - result.attrs_.set(index, builderForValue.build()); - return this; - } - public Builder addAttrs(org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute value) { - if (value == null) { - throw new NullPointerException(); - } - if (result.attrs_.isEmpty()) { - result.attrs_ = new java.util.ArrayList(); - } - result.attrs_.add(value); - return this; - } - public Builder addAttrs(org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute.Builder builderForValue) { - if (result.attrs_.isEmpty()) { - result.attrs_ = new java.util.ArrayList(); - } - result.attrs_.add(builderForValue.build()); - return this; - } - public Builder addAllAttrs( - java.lang.Iterable values) { - if (result.attrs_.isEmpty()) { - result.attrs_ = new java.util.ArrayList(); - } - super.addAll(values, result.attrs_); - return this; - } - public Builder clearAttrs() { - result.attrs_ = java.util.Collections.emptyList(); - return this; - } - - // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchema columns = 3; - public java.util.List getColumnsList() { - return java.util.Collections.unmodifiableList(result.columns_); - } - public int getColumnsCount() { - return result.getColumnsCount(); - } - public org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema getColumns(int index) { - return result.getColumns(index); - } - public Builder setColumns(int index, org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema value) { - if (value == null) { - throw new NullPointerException(); - } - result.columns_.set(index, value); - return this; - } - public Builder setColumns(int index, org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Builder builderForValue) { - result.columns_.set(index, builderForValue.build()); - return this; - } - public Builder addColumns(org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema value) { - if (value == null) { - throw new NullPointerException(); - } - if (result.columns_.isEmpty()) { - result.columns_ = new java.util.ArrayList(); - } - result.columns_.add(value); - return this; - } - public Builder addColumns(org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.ColumnSchema.Builder builderForValue) { - if (result.columns_.isEmpty()) { - result.columns_ = new java.util.ArrayList(); - } - result.columns_.add(builderForValue.build()); - return this; - } - public Builder addAllColumns( - java.lang.Iterable values) { - if (result.columns_.isEmpty()) { - result.columns_ = new java.util.ArrayList(); - } - super.addAll(values, result.columns_); - return this; - } - public Builder clearColumns() { - result.columns_ = java.util.Collections.emptyList(); - return this; - } - - // optional bool inMemory = 4; - public boolean hasInMemory() { - return result.hasInMemory(); - } - public boolean getInMemory() { - return result.getInMemory(); - } - public Builder setInMemory(boolean value) { - result.hasInMemory = true; - result.inMemory_ = value; - return this; - } - public Builder clearInMemory() { - result.hasInMemory = false; - result.inMemory_ = false; - return this; - } - - // optional bool readOnly = 5; - public boolean hasReadOnly() { - return result.hasReadOnly(); - } - public boolean getReadOnly() { - return result.getReadOnly(); - } - public Builder setReadOnly(boolean value) { - result.hasReadOnly = true; - result.readOnly_ = value; - return this; - } - public Builder clearReadOnly() { - result.hasReadOnly = false; - result.readOnly_ = false; - return this; - } - - // @@protoc_insertion_point(builder_scope:org.apache.hadoop.hbase.rest.protobuf.generated.TableSchema) - } - - static { - defaultInstance = new TableSchema(true); - org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.internalForceInit(); - defaultInstance.initFields(); - } - - // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.TableSchema) - } - - private static com.google.protobuf.Descriptors.Descriptor - internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable - internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor - internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_Attribute_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable - internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_Attribute_fieldAccessorTable; - - public static com.google.protobuf.Descriptors.FileDescriptor - getDescriptor() { - return descriptor; - } - private static com.google.protobuf.Descriptors.FileDescriptor - descriptor; - static { - java.lang.String[] descriptorData = { - "\n\030TableSchemaMessage.proto\022/org.apache.h" + - "adoop.hbase.rest.protobuf.generated\032\031Col" + - "umnSchemaMessage.proto\"\220\002\n\013TableSchema\022\014" + - "\n\004name\030\001 \001(\t\022U\n\005attrs\030\002 \003(\0132F.org.apache" + - ".hadoop.hbase.rest.protobuf.generated.Ta" + - "bleSchema.Attribute\022N\n\007columns\030\003 \003(\0132=.o" + - "rg.apache.hadoop.hbase.rest.protobuf.gen" + - "erated.ColumnSchema\022\020\n\010inMemory\030\004 \001(\010\022\020\n" + - "\010readOnly\030\005 \001(\010\032(\n\tAttribute\022\014\n\004name\030\001 \002" + - "(\t\022\r\n\005value\030\002 \002(\t" - }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { - descriptor = root; - internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_descriptor = - getDescriptor().getMessageTypes().get(0); - internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_descriptor, - new java.lang.String[] { "Name", "Attrs", "Columns", "InMemory", "ReadOnly", }, - org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.class, - org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Builder.class); - internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_Attribute_descriptor = - internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_descriptor.getNestedTypes().get(0); - internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_Attribute_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_TableSchema_Attribute_descriptor, - new java.lang.String[] { "Name", "Value", }, - org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute.class, - org.apache.hadoop.hbase.rest.protobuf.generated.TableSchemaMessage.TableSchema.Attribute.Builder.class); - return null; - } - }; - com.google.protobuf.Descriptors.FileDescriptor - .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { - org.apache.hadoop.hbase.rest.protobuf.generated.ColumnSchemaMessage.getDescriptor(), - }, assigner); - } - - public static void internalForceInit() {} - - // @@protoc_insertion_point(outer_class_scope) -} Modified: hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/VersionMessage.java URL: http://svn.apache.org/viewvc/hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/VersionMessage.java?rev=1594423&r1=1594422&r2=1594423&view=diff ============================================================================== --- hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/VersionMessage.java (original) +++ hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/VersionMessage.java Wed May 14 00:26:57 2014 @@ -1,511 +0,0 @@ -// Generated by the protocol buffer compiler. DO NOT EDIT! -// source: VersionMessage.proto - -package org.apache.hadoop.hbase.rest.protobuf.generated; - -public final class VersionMessage { - private VersionMessage() {} - public static void registerAllExtensions( - com.google.protobuf.ExtensionRegistry registry) { - } - public static final class Version extends - com.google.protobuf.GeneratedMessage { - // Use Version.newBuilder() to construct. - private Version() { - initFields(); - } - private Version(boolean noInit) {} - - private static final Version defaultInstance; - public static Version getDefaultInstance() { - return defaultInstance; - } - - public Version getDefaultInstanceForType() { - return defaultInstance; - } - - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Version_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Version_fieldAccessorTable; - } - - // optional string restVersion = 1; - public static final int RESTVERSION_FIELD_NUMBER = 1; - private boolean hasRestVersion; - private java.lang.String restVersion_ = ""; - public boolean hasRestVersion() { return hasRestVersion; } - public java.lang.String getRestVersion() { return restVersion_; } - - // optional string jvmVersion = 2; - public static final int JVMVERSION_FIELD_NUMBER = 2; - private boolean hasJvmVersion; - private java.lang.String jvmVersion_ = ""; - public boolean hasJvmVersion() { return hasJvmVersion; } - public java.lang.String getJvmVersion() { return jvmVersion_; } - - // optional string osVersion = 3; - public static final int OSVERSION_FIELD_NUMBER = 3; - private boolean hasOsVersion; - private java.lang.String osVersion_ = ""; - public boolean hasOsVersion() { return hasOsVersion; } - public java.lang.String getOsVersion() { return osVersion_; } - - // optional string serverVersion = 4; - public static final int SERVERVERSION_FIELD_NUMBER = 4; - private boolean hasServerVersion; - private java.lang.String serverVersion_ = ""; - public boolean hasServerVersion() { return hasServerVersion; } - public java.lang.String getServerVersion() { return serverVersion_; } - - // optional string jerseyVersion = 5; - public static final int JERSEYVERSION_FIELD_NUMBER = 5; - private boolean hasJerseyVersion; - private java.lang.String jerseyVersion_ = ""; - public boolean hasJerseyVersion() { return hasJerseyVersion; } - public java.lang.String getJerseyVersion() { return jerseyVersion_; } - - private void initFields() { - } - public final boolean isInitialized() { - return true; - } - - public void writeTo(com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - getSerializedSize(); - if (hasRestVersion()) { - output.writeString(1, getRestVersion()); - } - if (hasJvmVersion()) { - output.writeString(2, getJvmVersion()); - } - if (hasOsVersion()) { - output.writeString(3, getOsVersion()); - } - if (hasServerVersion()) { - output.writeString(4, getServerVersion()); - } - if (hasJerseyVersion()) { - output.writeString(5, getJerseyVersion()); - } - getUnknownFields().writeTo(output); - } - - private int memoizedSerializedSize = -1; - public int getSerializedSize() { - int size = memoizedSerializedSize; - if (size != -1) return size; - - size = 0; - if (hasRestVersion()) { - size += com.google.protobuf.CodedOutputStream - .computeStringSize(1, getRestVersion()); - } - if (hasJvmVersion()) { - size += com.google.protobuf.CodedOutputStream - .computeStringSize(2, getJvmVersion()); - } - if (hasOsVersion()) { - size += com.google.protobuf.CodedOutputStream - .computeStringSize(3, getOsVersion()); - } - if (hasServerVersion()) { - size += com.google.protobuf.CodedOutputStream - .computeStringSize(4, getServerVersion()); - } - if (hasJerseyVersion()) { - size += com.google.protobuf.CodedOutputStream - .computeStringSize(5, getJerseyVersion()); - } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; - return size; - } - - public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data).buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data).buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseFrom( - byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return newBuilder().mergeFrom(data, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseFrom(java.io.InputStream input) - throws java.io.IOException { - return newBuilder().mergeFrom(input).buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return newBuilder().mergeFrom(input, extensionRegistry) - .buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - Builder builder = newBuilder(); - if (builder.mergeDelimitedFrom(input)) { - return builder.buildParsed(); - } else { - return null; - } - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseDelimitedFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - Builder builder = newBuilder(); - if (builder.mergeDelimitedFrom(input, extensionRegistry)) { - return builder.buildParsed(); - } else { - return null; - } - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseFrom( - com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return newBuilder().mergeFrom(input).buildParsed(); - } - public static org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return newBuilder().mergeFrom(input, extensionRegistry) - .buildParsed(); - } - - public static Builder newBuilder() { return Builder.create(); } - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version prototype) { - return newBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { return newBuilder(this); } - - public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder { - private org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version result; - - // Construct using org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version.newBuilder() - private Builder() {} - - private static Builder create() { - Builder builder = new Builder(); - builder.result = new org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version(); - return builder; - } - - protected org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version internalGetResult() { - return result; - } - - public Builder clear() { - if (result == null) { - throw new IllegalStateException( - "Cannot call clear() after build()."); - } - result = new org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version(); - return this; - } - - public Builder clone() { - return create().mergeFrom(result); - } - - public com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version.getDescriptor(); - } - - public org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version getDefaultInstanceForType() { - return org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version.getDefaultInstance(); - } - - public boolean isInitialized() { - return result.isInitialized(); - } - public org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version build() { - if (result != null && !isInitialized()) { - throw newUninitializedMessageException(result); - } - return buildPartial(); - } - - private org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version buildParsed() - throws com.google.protobuf.InvalidProtocolBufferException { - if (!isInitialized()) { - throw newUninitializedMessageException( - result).asInvalidProtocolBufferException(); - } - return buildPartial(); - } - - public org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version buildPartial() { - if (result == null) { - throw new IllegalStateException( - "build() has already been called on this Builder."); - } - org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version returnMe = result; - result = null; - return returnMe; - } - - public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version) { - return mergeFrom((org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version other) { - if (other == org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version.getDefaultInstance()) return this; - if (other.hasRestVersion()) { - setRestVersion(other.getRestVersion()); - } - if (other.hasJvmVersion()) { - setJvmVersion(other.getJvmVersion()); - } - if (other.hasOsVersion()) { - setOsVersion(other.getOsVersion()); - } - if (other.hasServerVersion()) { - setServerVersion(other.getServerVersion()); - } - if (other.hasJerseyVersion()) { - setJerseyVersion(other.getJerseyVersion()); - } - this.mergeUnknownFields(other.getUnknownFields()); - return this; - } - - public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder( - this.getUnknownFields()); - while (true) { - int tag = input.readTag(); - switch (tag) { - case 0: - this.setUnknownFields(unknownFields.build()); - return this; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - this.setUnknownFields(unknownFields.build()); - return this; - } - break; - } - case 10: { - setRestVersion(input.readString()); - break; - } - case 18: { - setJvmVersion(input.readString()); - break; - } - case 26: { - setOsVersion(input.readString()); - break; - } - case 34: { - setServerVersion(input.readString()); - break; - } - case 42: { - setJerseyVersion(input.readString()); - break; - } - } - } - } - - - // optional string restVersion = 1; - public boolean hasRestVersion() { - return result.hasRestVersion(); - } - public java.lang.String getRestVersion() { - return result.getRestVersion(); - } - public Builder setRestVersion(java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - result.hasRestVersion = true; - result.restVersion_ = value; - return this; - } - public Builder clearRestVersion() { - result.hasRestVersion = false; - result.restVersion_ = getDefaultInstance().getRestVersion(); - return this; - } - - // optional string jvmVersion = 2; - public boolean hasJvmVersion() { - return result.hasJvmVersion(); - } - public java.lang.String getJvmVersion() { - return result.getJvmVersion(); - } - public Builder setJvmVersion(java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - result.hasJvmVersion = true; - result.jvmVersion_ = value; - return this; - } - public Builder clearJvmVersion() { - result.hasJvmVersion = false; - result.jvmVersion_ = getDefaultInstance().getJvmVersion(); - return this; - } - - // optional string osVersion = 3; - public boolean hasOsVersion() { - return result.hasOsVersion(); - } - public java.lang.String getOsVersion() { - return result.getOsVersion(); - } - public Builder setOsVersion(java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - result.hasOsVersion = true; - result.osVersion_ = value; - return this; - } - public Builder clearOsVersion() { - result.hasOsVersion = false; - result.osVersion_ = getDefaultInstance().getOsVersion(); - return this; - } - - // optional string serverVersion = 4; - public boolean hasServerVersion() { - return result.hasServerVersion(); - } - public java.lang.String getServerVersion() { - return result.getServerVersion(); - } - public Builder setServerVersion(java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - result.hasServerVersion = true; - result.serverVersion_ = value; - return this; - } - public Builder clearServerVersion() { - result.hasServerVersion = false; - result.serverVersion_ = getDefaultInstance().getServerVersion(); - return this; - } - - // optional string jerseyVersion = 5; - public boolean hasJerseyVersion() { - return result.hasJerseyVersion(); - } - public java.lang.String getJerseyVersion() { - return result.getJerseyVersion(); - } - public Builder setJerseyVersion(java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - result.hasJerseyVersion = true; - result.jerseyVersion_ = value; - return this; - } - public Builder clearJerseyVersion() { - result.hasJerseyVersion = false; - result.jerseyVersion_ = getDefaultInstance().getJerseyVersion(); - return this; - } - - // @@protoc_insertion_point(builder_scope:org.apache.hadoop.hbase.rest.protobuf.generated.Version) - } - - static { - defaultInstance = new Version(true); - org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.internalForceInit(); - defaultInstance.initFields(); - } - - // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.Version) - } - - private static com.google.protobuf.Descriptors.Descriptor - internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Version_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable - internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Version_fieldAccessorTable; - - public static com.google.protobuf.Descriptors.FileDescriptor - getDescriptor() { - return descriptor; - } - private static com.google.protobuf.Descriptors.FileDescriptor - descriptor; - static { - java.lang.String[] descriptorData = { - "\n\024VersionMessage.proto\022/org.apache.hadoo" + - "p.hbase.rest.protobuf.generated\"s\n\007Versi" + - "on\022\023\n\013restVersion\030\001 \001(\t\022\022\n\njvmVersion\030\002 " + - "\001(\t\022\021\n\tosVersion\030\003 \001(\t\022\025\n\rserverVersion\030" + - "\004 \001(\t\022\025\n\rjerseyVersion\030\005 \001(\t" - }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { - descriptor = root; - internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Version_descriptor = - getDescriptor().getMessageTypes().get(0); - internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Version_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_Version_descriptor, - new java.lang.String[] { "RestVersion", "JvmVersion", "OsVersion", "ServerVersion", "JerseyVersion", }, - org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version.class, - org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version.Builder.class); - return null; - } - }; - com.google.protobuf.Descriptors.FileDescriptor - .internalBuildGeneratedFileFrom(descriptorData, - new com.google.protobuf.Descriptors.FileDescriptor[] { - }, assigner); - } - - public static void internalForceInit() {} - - // @@protoc_insertion_point(outer_class_scope) -} Modified: hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java URL: http://svn.apache.org/viewvc/hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java?rev=1594423&r1=1594422&r2=1594423&view=diff ============================================================================== --- hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java (original) +++ hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java Wed May 14 00:26:57 2014 @@ -1,88 +0,0 @@ -/* - * Copyright 2010 The Apache Software Foundation - * - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -package org.apache.hadoop.hbase.rest.provider; - -import java.util.Arrays; -import java.util.HashSet; -import java.util.Set; - -import javax.ws.rs.ext.ContextResolver; -import javax.ws.rs.ext.Provider; -import javax.xml.bind.JAXBContext; - -import org.apache.hadoop.hbase.rest.model.CellModel; -import org.apache.hadoop.hbase.rest.model.CellSetModel; -import org.apache.hadoop.hbase.rest.model.ColumnSchemaModel; -import org.apache.hadoop.hbase.rest.model.RowModel; -import org.apache.hadoop.hbase.rest.model.ScannerModel; -import org.apache.hadoop.hbase.rest.model.StorageClusterStatusModel; -import org.apache.hadoop.hbase.rest.model.StorageClusterVersionModel; -import org.apache.hadoop.hbase.rest.model.TableInfoModel; -import org.apache.hadoop.hbase.rest.model.TableListModel; -import org.apache.hadoop.hbase.rest.model.TableModel; -import org.apache.hadoop.hbase.rest.model.TableRegionModel; -import org.apache.hadoop.hbase.rest.model.TableSchemaModel; -import org.apache.hadoop.hbase.rest.model.VersionModel; - -import com.sun.jersey.api.json.JSONConfiguration; -import com.sun.jersey.api.json.JSONJAXBContext; - -/** - * Plumbing for hooking up Jersey's JSON entity body encoding and decoding - * support to JAXB. Modify how the context is created (by using e.g. a - * different configuration builder) to control how JSON is processed and - * created. - */ -@Provider -public class JAXBContextResolver implements ContextResolver { - - private final JAXBContext context; - - private final Set> types; - - private final Class[] cTypes = { - CellModel.class, - CellSetModel.class, - ColumnSchemaModel.class, - RowModel.class, - ScannerModel.class, - StorageClusterStatusModel.class, - StorageClusterVersionModel.class, - TableInfoModel.class, - TableListModel.class, - TableModel.class, - TableRegionModel.class, - TableSchemaModel.class, - VersionModel.class - }; - - @SuppressWarnings("unchecked") - public JAXBContextResolver() throws Exception { - this.types = new HashSet(Arrays.asList(cTypes)); - this.context = new JSONJAXBContext(JSONConfiguration.natural().build(), - cTypes); - } - - @Override - public JAXBContext getContext(Class objectType) { - return (types.contains(objectType)) ? context : null; - } -} Modified: hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/provider/consumer/ProtobufMessageBodyConsumer.java URL: http://svn.apache.org/viewvc/hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/provider/consumer/ProtobufMessageBodyConsumer.java?rev=1594423&r1=1594422&r2=1594423&view=diff ============================================================================== --- hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/provider/consumer/ProtobufMessageBodyConsumer.java (original) +++ hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/provider/consumer/ProtobufMessageBodyConsumer.java Wed May 14 00:26:57 2014 @@ -1,87 +0,0 @@ -/* - * Copyright 2010 The Apache Software Foundation - * - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -package org.apache.hadoop.hbase.rest.provider.consumer; - -import java.io.ByteArrayOutputStream; -import java.io.IOException; -import java.io.InputStream; -import java.lang.annotation.Annotation; -import java.lang.reflect.Type; - -import javax.ws.rs.Consumes; -import javax.ws.rs.WebApplicationException; -import javax.ws.rs.core.MediaType; -import javax.ws.rs.core.MultivaluedMap; -import javax.ws.rs.ext.MessageBodyReader; -import javax.ws.rs.ext.Provider; - -import org.apache.commons.logging.Log; -import org.apache.commons.logging.LogFactory; -import org.apache.hadoop.hbase.rest.Constants; -import org.apache.hadoop.hbase.rest.ProtobufMessageHandler; - -/** - * Adapter for hooking up Jersey content processing dispatch to - * ProtobufMessageHandler interface capable handlers for decoding protobuf input. - */ -@Provider -@Consumes(Constants.MIMETYPE_PROTOBUF) -public class ProtobufMessageBodyConsumer - implements MessageBodyReader { - private static final Log LOG = - LogFactory.getLog(ProtobufMessageBodyConsumer.class); - - @Override - public boolean isReadable(Class type, Type genericType, - Annotation[] annotations, MediaType mediaType) { - return ProtobufMessageHandler.class.isAssignableFrom(type); - } - - @Override - public ProtobufMessageHandler readFrom(Class type, Type genericType, - Annotation[] annotations, MediaType mediaType, - MultivaluedMap httpHeaders, InputStream inputStream) - throws IOException, WebApplicationException { - ProtobufMessageHandler obj = null; - try { - obj = type.newInstance(); - ByteArrayOutputStream baos = new ByteArrayOutputStream(); - byte[] buffer = new byte[4096]; - int read; - do { - read = inputStream.read(buffer, 0, buffer.length); - if (read > 0) { - baos.write(buffer, 0, read); - } - } while (read > 0); - if (LOG.isDebugEnabled()) { - LOG.debug(getClass() + ": read " + baos.size() + " bytes from " + - inputStream); - } - obj = obj.getObjectFromMessage(baos.toByteArray()); - } catch (InstantiationException e) { - throw new WebApplicationException(e); - } catch (IllegalAccessException e) { - throw new WebApplicationException(e); - } - return obj; - } -} Modified: hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/provider/producer/PlainTextMessageBodyProducer.java URL: http://svn.apache.org/viewvc/hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/provider/producer/PlainTextMessageBodyProducer.java?rev=1594423&r1=1594422&r2=1594423&view=diff ============================================================================== --- hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/provider/producer/PlainTextMessageBodyProducer.java (original) +++ hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/provider/producer/PlainTextMessageBodyProducer.java Wed May 14 00:26:57 2014 @@ -1,73 +0,0 @@ -/* - * Copyright 2010 The Apache Software Foundation - * - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -package org.apache.hadoop.hbase.rest.provider.producer; - -import java.io.IOException; -import java.io.OutputStream; -import java.lang.annotation.Annotation; -import java.lang.reflect.Type; -import java.util.Map; -import java.util.WeakHashMap; - -import javax.ws.rs.Produces; -import javax.ws.rs.WebApplicationException; -import javax.ws.rs.core.MediaType; -import javax.ws.rs.core.MultivaluedMap; -import javax.ws.rs.ext.MessageBodyWriter; -import javax.ws.rs.ext.Provider; - -import org.apache.hadoop.hbase.rest.Constants; - -/** - * An adapter between Jersey and Object.toString(). Hooks up plain text output - * to the Jersey content handling framework. - * Jersey will first call getSize() to learn the number of bytes that will be - * sent, then writeTo to perform the actual I/O. - */ -@Provider -@Produces(Constants.MIMETYPE_TEXT) -public class PlainTextMessageBodyProducer - implements MessageBodyWriter { - - private Map buffer = new WeakHashMap(); - - @Override - public boolean isWriteable(Class arg0, Type arg1, Annotation[] arg2, - MediaType arg3) { - return true; - } - - @Override - public long getSize(Object object, Class type, Type genericType, - Annotation[] annotations, MediaType mediaType) { - byte[] bytes = object.toString().getBytes(); - buffer.put(object, bytes); - return bytes.length; - } - - @Override - public void writeTo(Object object, Class type, Type genericType, - Annotation[] annotations, MediaType mediaType, - MultivaluedMap httpHeaders, OutputStream outStream) - throws IOException, WebApplicationException { - outStream.write(buffer.remove(object)); - } -} Modified: hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/provider/producer/ProtobufMessageBodyProducer.java URL: http://svn.apache.org/viewvc/hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/provider/producer/ProtobufMessageBodyProducer.java?rev=1594423&r1=1594422&r2=1594423&view=diff ============================================================================== --- hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/provider/producer/ProtobufMessageBodyProducer.java (original) +++ hbase/branches/0.89-fb/src/main/java/org/apache/hadoop/hbase/rest/provider/producer/ProtobufMessageBodyProducer.java Wed May 14 00:26:57 2014 @@ -1,80 +0,0 @@ -/* - * Copyright 2010 The Apache Software Foundation - * - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -package org.apache.hadoop.hbase.rest.provider.producer; - -import java.io.ByteArrayOutputStream; -import java.io.IOException; -import java.io.OutputStream; -import java.lang.annotation.Annotation; -import java.lang.reflect.Type; -import java.util.Map; -import java.util.WeakHashMap; - -import javax.ws.rs.Produces; -import javax.ws.rs.WebApplicationException; -import javax.ws.rs.core.MediaType; -import javax.ws.rs.core.MultivaluedMap; -import javax.ws.rs.ext.MessageBodyWriter; -import javax.ws.rs.ext.Provider; - -import org.apache.hadoop.hbase.rest.Constants; -import org.apache.hadoop.hbase.rest.ProtobufMessageHandler; - -/** - * An adapter between Jersey and ProtobufMessageHandler implementors. Hooks up - * protobuf output producing methods to the Jersey content handling framework. - * Jersey will first call getSize() to learn the number of bytes that will be - * sent, then writeTo to perform the actual I/O. - */ -@Provider -@Produces(Constants.MIMETYPE_PROTOBUF) -public class ProtobufMessageBodyProducer - implements MessageBodyWriter { - - private Map buffer = new WeakHashMap(); - - @Override - public boolean isWriteable(Class type, Type genericType, - Annotation[] annotations, MediaType mediaType) { - return ProtobufMessageHandler.class.isAssignableFrom(type); - } - - @Override - public long getSize(ProtobufMessageHandler m, Class type, Type genericType, - Annotation[] annotations, MediaType mediaType) { - ByteArrayOutputStream baos = new ByteArrayOutputStream(); - try { - baos.write(m.createProtobufOutput()); - } catch (IOException e) { - return -1; - } - byte[] bytes = baos.toByteArray(); - buffer.put(m, bytes); - return bytes.length; - } - - public void writeTo(ProtobufMessageHandler m, Class type, Type genericType, - Annotation[] annotations, MediaType mediaType, - MultivaluedMap httpHeaders, OutputStream entityStream) - throws IOException, WebApplicationException { - entityStream.write(buffer.remove(m)); - } -} Modified: hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/HFilePerformanceEvaluation.java URL: http://svn.apache.org/viewvc/hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/HFilePerformanceEvaluation.java?rev=1594423&r1=1594422&r2=1594423&view=diff ============================================================================== --- hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/HFilePerformanceEvaluation.java (original) +++ hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/HFilePerformanceEvaluation.java Wed May 14 00:26:57 2014 @@ -368,4 +368,4 @@ public class HFilePerformanceEvaluation public static void main(String[] args) throws Exception { new HFilePerformanceEvaluation().runBenchmarks(); } -} \ No newline at end of file +} Modified: hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/MultiRegionTable.java URL: http://svn.apache.org/viewvc/hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/MultiRegionTable.java?rev=1594423&r1=1594422&r2=1594423&view=diff ============================================================================== --- hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/MultiRegionTable.java (original) +++ hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/MultiRegionTable.java Wed May 14 00:26:57 2014 @@ -115,4 +115,4 @@ public class MultiRegionTable extends HB region.close(); region.getLog().closeAndDelete(); } -} \ No newline at end of file +} Modified: hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluationCommons.java URL: http://svn.apache.org/viewvc/hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluationCommons.java?rev=1594423&r1=1594422&r2=1594423&view=diff ============================================================================== --- hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluationCommons.java (original) +++ hbase/branches/0.89-fb/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluationCommons.java Wed May 14 00:26:57 2014 @@ -75,4 +75,4 @@ public class PerformanceEvaluationCommon } LOG.info("Test took " + (System.currentTimeMillis() - now)); } -} \ No newline at end of file +}