Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id EFB742004A0 for ; Wed, 16 Aug 2017 20:35:37 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id EE2D5169535; Wed, 16 Aug 2017 18:35:37 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 7E485169537 for ; Wed, 16 Aug 2017 20:35:34 +0200 (CEST) Received: (qmail 21468 invoked by uid 500); 16 Aug 2017 18:35:33 -0000 Mailing-List: contact commits-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@hbase.apache.org Delivered-To: mailing list commits@hbase.apache.org Received: (qmail 21248 invoked by uid 99); 16 Aug 2017 18:35:33 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 16 Aug 2017 18:35:33 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id DFB2DF32B3; Wed, 16 Aug 2017 18:35:32 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: apurtell@apache.org To: commits@hbase.apache.org Date: Wed, 16 Aug 2017 18:35:34 -0000 Message-Id: <69d9c55c031243e3a99a0b2a2b63ec77@git.apache.org> In-Reply-To: <8ee007c4dc844ec9be3bd4faf0104842@git.apache.org> References: <8ee007c4dc844ec9be3bd4faf0104842@git.apache.org> X-Mailer: ASF-Git Admin Mailer Subject: [03/14] hbase git commit: HBASE-18431 Mitigate compatibility concerns between branch-1.3 and branch-1.4 archived-at: Wed, 16 Aug 2017 18:35:38 -0000 http://git-wip-us.apache.org/repos/asf/hbase/blob/ea5789a7/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SecureBulkLoadProtos.java ---------------------------------------------------------------------- diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SecureBulkLoadProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SecureBulkLoadProtos.java index 538e031..8521ba8 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SecureBulkLoadProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SecureBulkLoadProtos.java @@ -2517,11 +2517,11 @@ public final class SecureBulkLoadProtos { /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder(); } /** * Protobuf type {@code hbase.pb.PrepareBulkLoadRequest} @@ -2575,11 +2575,11 @@ public final class SecureBulkLoadProtos { break; } case 10: { - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = tableName_.toBuilder(); } - tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); + tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); @@ -2629,7 +2629,7 @@ public final class SecureBulkLoadProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -2639,18 +2639,18 @@ public final class SecureBulkLoadProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { return tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } private void initFields() { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -2844,7 +2844,7 @@ public final class SecureBulkLoadProtos { public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } @@ -2940,9 +2940,9 @@ public final class SecureBulkLoadProtos { private int bitField0_; // required .hbase.pb.TableName table_name = 1; - private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + private org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; */ @@ -2952,7 +2952,7 @@ public final class SecureBulkLoadProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { @@ -2962,7 +2962,7 @@ public final class SecureBulkLoadProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); @@ -2979,7 +2979,7 @@ public final class SecureBulkLoadProtos { * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); @@ -2992,12 +2992,12 @@ public final class SecureBulkLoadProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { + public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && - tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { + tableName_ != org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance()) { tableName_ = - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } @@ -3013,7 +3013,7 @@ public final class SecureBulkLoadProtos { */ public Builder clearTableName() { if (tableNameBuilder_ == null) { - tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); + tableName_ = org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); @@ -3024,7 +3024,7 @@ public final class SecureBulkLoadProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); @@ -3032,7 +3032,7 @@ public final class SecureBulkLoadProtos { /** * required .hbase.pb.TableName table_name = 1; */ - public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { + public org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { @@ -3043,11 +3043,11 @@ public final class SecureBulkLoadProtos { * required .hbase.pb.TableName table_name = 1; */ private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( + org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); @@ -4857,30 +4857,31 @@ public final class SecureBulkLoadProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\024SecureBulkLoad.proto\022\010hbase.pb\032\013HBase." + - "proto\032\014Client.proto\"\266\001\n\033SecureBulkLoadHF" + - "ilesRequest\022>\n\013family_path\030\001 \003(\0132).hbase" + - ".pb.BulkLoadHFileRequest.FamilyPath\022\026\n\016a" + - "ssign_seq_num\030\002 \001(\010\022+\n\010fs_token\030\003 \002(\0132\031." + - "hbase.pb.DelegationToken\022\022\n\nbulk_token\030\004" + - " \002(\t\".\n\034SecureBulkLoadHFilesResponse\022\016\n\006" + - "loaded\030\001 \002(\010\"V\n\017DelegationToken\022\022\n\nident" + - "ifier\030\001 \001(\014\022\020\n\010password\030\002 \001(\014\022\014\n\004kind\030\003 " + - "\001(\t\022\017\n\007service\030\004 \001(\t\"A\n\026PrepareBulkLoadR", - "equest\022\'\n\ntable_name\030\001 \002(\0132\023.hbase.pb.Ta" + - "bleName\"-\n\027PrepareBulkLoadResponse\022\022\n\nbu" + - "lk_token\030\001 \002(\t\",\n\026CleanupBulkLoadRequest" + - "\022\022\n\nbulk_token\030\001 \002(\t\"\031\n\027CleanupBulkLoadR" + - "esponse2\256\002\n\025SecureBulkLoadService\022V\n\017Pre" + - "pareBulkLoad\022 .hbase.pb.PrepareBulkLoadR" + - "equest\032!.hbase.pb.PrepareBulkLoadRespons" + - "e\022e\n\024SecureBulkLoadHFiles\022%.hbase.pb.Sec" + - "ureBulkLoadHFilesRequest\032&.hbase.pb.Secu" + - "reBulkLoadHFilesResponse\022V\n\017CleanupBulkL", - "oad\022 .hbase.pb.CleanupBulkLoadRequest\032!." + - "hbase.pb.CleanupBulkLoadResponseBJ\n*org." + - "apache.hadoop.hbase.protobuf.generatedB\024" + - "SecureBulkLoadProtosH\001\210\001\001\240\001\001" + "\n\024SecureBulkLoad.proto\022\010hbase.pb\032\013Table." + + "proto\032\013HBase.proto\032\014Client.proto\"\266\001\n\033Sec" + + "ureBulkLoadHFilesRequest\022>\n\013family_path\030" + + "\001 \003(\0132).hbase.pb.BulkLoadHFileRequest.Fa" + + "milyPath\022\026\n\016assign_seq_num\030\002 \001(\010\022+\n\010fs_t" + + "oken\030\003 \002(\0132\031.hbase.pb.DelegationToken\022\022\n" + + "\nbulk_token\030\004 \002(\t\".\n\034SecureBulkLoadHFile" + + "sResponse\022\016\n\006loaded\030\001 \002(\010\"V\n\017DelegationT" + + "oken\022\022\n\nidentifier\030\001 \001(\014\022\020\n\010password\030\002 \001" + + "(\014\022\014\n\004kind\030\003 \001(\t\022\017\n\007service\030\004 \001(\t\"A\n\026Pre", + "pareBulkLoadRequest\022\'\n\ntable_name\030\001 \002(\0132" + + "\023.hbase.pb.TableName\"-\n\027PrepareBulkLoadR" + + "esponse\022\022\n\nbulk_token\030\001 \002(\t\",\n\026CleanupBu" + + "lkLoadRequest\022\022\n\nbulk_token\030\001 \002(\t\"\031\n\027Cle" + + "anupBulkLoadResponse2\256\002\n\025SecureBulkLoadS" + + "ervice\022V\n\017PrepareBulkLoad\022 .hbase.pb.Pre" + + "pareBulkLoadRequest\032!.hbase.pb.PrepareBu" + + "lkLoadResponse\022e\n\024SecureBulkLoadHFiles\022%" + + ".hbase.pb.SecureBulkLoadHFilesRequest\032&." + + "hbase.pb.SecureBulkLoadHFilesResponse\022V\n", + "\017CleanupBulkLoad\022 .hbase.pb.CleanupBulkL" + + "oadRequest\032!.hbase.pb.CleanupBulkLoadRes" + + "ponseBJ\n*org.apache.hadoop.hbase.protobu" + + "f.generatedB\024SecureBulkLoadProtosH\001\210\001\001\240\001" + + "\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { @@ -4935,6 +4936,7 @@ public final class SecureBulkLoadProtos { com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { + org.apache.hadoop.hbase.protobuf.generated.TableProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.ClientProtos.getDescriptor(), }, assigner); http://git-wip-us.apache.org/repos/asf/hbase/blob/ea5789a7/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SnapshotProtos.java ---------------------------------------------------------------------- diff --git a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SnapshotProtos.java b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SnapshotProtos.java index 2e11b4a..9805d50 100644 --- a/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SnapshotProtos.java +++ b/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SnapshotProtos.java @@ -8,1494 +8,6 @@ public final class SnapshotProtos { public static void registerAllExtensions( com.google.protobuf.ExtensionRegistry registry) { } - public interface SnapshotDescriptionOrBuilder - extends com.google.protobuf.MessageOrBuilder { - - // required string name = 1; - /** - * required string name = 1; - */ - boolean hasName(); - /** - * required string name = 1; - */ - java.lang.String getName(); - /** - * required string name = 1; - */ - com.google.protobuf.ByteString - getNameBytes(); - - // optional string table = 2; - /** - * optional string table = 2; - * - *
-     * not needed for delete, but checked for in taking snapshot
-     * 
- */ - boolean hasTable(); - /** - * optional string table = 2; - * - *
-     * not needed for delete, but checked for in taking snapshot
-     * 
- */ - java.lang.String getTable(); - /** - * optional string table = 2; - * - *
-     * not needed for delete, but checked for in taking snapshot
-     * 
- */ - com.google.protobuf.ByteString - getTableBytes(); - - // optional int64 creation_time = 3 [default = 0]; - /** - * optional int64 creation_time = 3 [default = 0]; - */ - boolean hasCreationTime(); - /** - * optional int64 creation_time = 3 [default = 0]; - */ - long getCreationTime(); - - // optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - /** - * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - */ - boolean hasType(); - /** - * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - */ - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type getType(); - - // optional int32 version = 5; - /** - * optional int32 version = 5; - */ - boolean hasVersion(); - /** - * optional int32 version = 5; - */ - int getVersion(); - - // optional string owner = 6; - /** - * optional string owner = 6; - */ - boolean hasOwner(); - /** - * optional string owner = 6; - */ - java.lang.String getOwner(); - /** - * optional string owner = 6; - */ - com.google.protobuf.ByteString - getOwnerBytes(); - - // optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - boolean hasUsersAndPermissions(); - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions getUsersAndPermissions(); - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissionsOrBuilder getUsersAndPermissionsOrBuilder(); - } - /** - * Protobuf type {@code hbase.pb.SnapshotDescription} - * - *
-   **
-   * Description of the snapshot to take
-   * 
- */ - public static final class SnapshotDescription extends - com.google.protobuf.GeneratedMessage - implements SnapshotDescriptionOrBuilder { - // Use SnapshotDescription.newBuilder() to construct. - private SnapshotDescription(com.google.protobuf.GeneratedMessage.Builder builder) { - super(builder); - this.unknownFields = builder.getUnknownFields(); - } - private SnapshotDescription(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final SnapshotDescription defaultInstance; - public static SnapshotDescription getDefaultInstance() { - return defaultInstance; - } - - public SnapshotDescription getDefaultInstanceForType() { - return defaultInstance; - } - - private final com.google.protobuf.UnknownFieldSet unknownFields; - @java.lang.Override - public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { - return this.unknownFields; - } - private SnapshotDescription( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); - int mutable_bitField0_ = 0; - com.google.protobuf.UnknownFieldSet.Builder unknownFields = - com.google.protobuf.UnknownFieldSet.newBuilder(); - try { - boolean done = false; - while (!done) { - int tag = input.readTag(); - switch (tag) { - case 0: - done = true; - break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } - case 10: { - bitField0_ |= 0x00000001; - name_ = input.readBytes(); - break; - } - case 18: { - bitField0_ |= 0x00000002; - table_ = input.readBytes(); - break; - } - case 24: { - bitField0_ |= 0x00000004; - creationTime_ = input.readInt64(); - break; - } - case 32: { - int rawValue = input.readEnum(); - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type value = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type.valueOf(rawValue); - if (value == null) { - unknownFields.mergeVarintField(4, rawValue); - } else { - bitField0_ |= 0x00000008; - type_ = value; - } - break; - } - case 40: { - bitField0_ |= 0x00000010; - version_ = input.readInt32(); - break; - } - case 50: { - bitField0_ |= 0x00000020; - owner_ = input.readBytes(); - break; - } - case 58: { - org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.Builder subBuilder = null; - if (((bitField0_ & 0x00000040) == 0x00000040)) { - subBuilder = usersAndPermissions_.toBuilder(); - } - usersAndPermissions_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.PARSER, extensionRegistry); - if (subBuilder != null) { - subBuilder.mergeFrom(usersAndPermissions_); - usersAndPermissions_ = subBuilder.buildPartial(); - } - bitField0_ |= 0x00000040; - break; - } - } - } - } catch (com.google.protobuf.InvalidProtocolBufferException e) { - throw e.setUnfinishedMessage(this); - } catch (java.io.IOException e) { - throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); - } finally { - this.unknownFields = unknownFields.build(); - makeExtensionsImmutable(); - } - } - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotDescription_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotDescription_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.class, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder.class); - } - - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public SnapshotDescription parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new SnapshotDescription(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - - /** - * Protobuf enum {@code hbase.pb.SnapshotDescription.Type} - */ - public enum Type - implements com.google.protobuf.ProtocolMessageEnum { - /** - * DISABLED = 0; - */ - DISABLED(0, 0), - /** - * FLUSH = 1; - */ - FLUSH(1, 1), - /** - * SKIPFLUSH = 2; - */ - SKIPFLUSH(2, 2), - ; - - /** - * DISABLED = 0; - */ - public static final int DISABLED_VALUE = 0; - /** - * FLUSH = 1; - */ - public static final int FLUSH_VALUE = 1; - /** - * SKIPFLUSH = 2; - */ - public static final int SKIPFLUSH_VALUE = 2; - - - public final int getNumber() { return value; } - - public static Type valueOf(int value) { - switch (value) { - case 0: return DISABLED; - case 1: return FLUSH; - case 2: return SKIPFLUSH; - default: return null; - } - } - - public static com.google.protobuf.Internal.EnumLiteMap - internalGetValueMap() { - return internalValueMap; - } - private static com.google.protobuf.Internal.EnumLiteMap - internalValueMap = - new com.google.protobuf.Internal.EnumLiteMap() { - public Type findValueByNumber(int number) { - return Type.valueOf(number); - } - }; - - public final com.google.protobuf.Descriptors.EnumValueDescriptor - getValueDescriptor() { - return getDescriptor().getValues().get(index); - } - public final com.google.protobuf.Descriptors.EnumDescriptor - getDescriptorForType() { - return getDescriptor(); - } - public static final com.google.protobuf.Descriptors.EnumDescriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDescriptor().getEnumTypes().get(0); - } - - private static final Type[] VALUES = values(); - - public static Type valueOf( - com.google.protobuf.Descriptors.EnumValueDescriptor desc) { - if (desc.getType() != getDescriptor()) { - throw new java.lang.IllegalArgumentException( - "EnumValueDescriptor is not for this type."); - } - return VALUES[desc.getIndex()]; - } - - private final int index; - private final int value; - - private Type(int index, int value) { - this.index = index; - this.value = value; - } - - // @@protoc_insertion_point(enum_scope:hbase.pb.SnapshotDescription.Type) - } - - private int bitField0_; - // required string name = 1; - public static final int NAME_FIELD_NUMBER = 1; - private java.lang.Object name_; - /** - * required string name = 1; - */ - public boolean hasName() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * required string name = 1; - */ - public java.lang.String getName() { - java.lang.Object ref = name_; - if (ref instanceof java.lang.String) { - return (java.lang.String) ref; - } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; - java.lang.String s = bs.toStringUtf8(); - if (bs.isValidUtf8()) { - name_ = s; - } - return s; - } - } - /** - * required string name = 1; - */ - public com.google.protobuf.ByteString - getNameBytes() { - java.lang.Object ref = name_; - if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - name_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - - // optional string table = 2; - public static final int TABLE_FIELD_NUMBER = 2; - private java.lang.Object table_; - /** - * optional string table = 2; - * - *
-     * not needed for delete, but checked for in taking snapshot
-     * 
- */ - public boolean hasTable() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * optional string table = 2; - * - *
-     * not needed for delete, but checked for in taking snapshot
-     * 
- */ - public java.lang.String getTable() { - java.lang.Object ref = table_; - if (ref instanceof java.lang.String) { - return (java.lang.String) ref; - } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; - java.lang.String s = bs.toStringUtf8(); - if (bs.isValidUtf8()) { - table_ = s; - } - return s; - } - } - /** - * optional string table = 2; - * - *
-     * not needed for delete, but checked for in taking snapshot
-     * 
- */ - public com.google.protobuf.ByteString - getTableBytes() { - java.lang.Object ref = table_; - if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - table_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - - // optional int64 creation_time = 3 [default = 0]; - public static final int CREATION_TIME_FIELD_NUMBER = 3; - private long creationTime_; - /** - * optional int64 creation_time = 3 [default = 0]; - */ - public boolean hasCreationTime() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - /** - * optional int64 creation_time = 3 [default = 0]; - */ - public long getCreationTime() { - return creationTime_; - } - - // optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - public static final int TYPE_FIELD_NUMBER = 4; - private org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type type_; - /** - * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - */ - public boolean hasType() { - return ((bitField0_ & 0x00000008) == 0x00000008); - } - /** - * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type getType() { - return type_; - } - - // optional int32 version = 5; - public static final int VERSION_FIELD_NUMBER = 5; - private int version_; - /** - * optional int32 version = 5; - */ - public boolean hasVersion() { - return ((bitField0_ & 0x00000010) == 0x00000010); - } - /** - * optional int32 version = 5; - */ - public int getVersion() { - return version_; - } - - // optional string owner = 6; - public static final int OWNER_FIELD_NUMBER = 6; - private java.lang.Object owner_; - /** - * optional string owner = 6; - */ - public boolean hasOwner() { - return ((bitField0_ & 0x00000020) == 0x00000020); - } - /** - * optional string owner = 6; - */ - public java.lang.String getOwner() { - java.lang.Object ref = owner_; - if (ref instanceof java.lang.String) { - return (java.lang.String) ref; - } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; - java.lang.String s = bs.toStringUtf8(); - if (bs.isValidUtf8()) { - owner_ = s; - } - return s; - } - } - /** - * optional string owner = 6; - */ - public com.google.protobuf.ByteString - getOwnerBytes() { - java.lang.Object ref = owner_; - if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - owner_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - - // optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - public static final int USERS_AND_PERMISSIONS_FIELD_NUMBER = 7; - private org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions usersAndPermissions_; - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public boolean hasUsersAndPermissions() { - return ((bitField0_ & 0x00000040) == 0x00000040); - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions getUsersAndPermissions() { - return usersAndPermissions_; - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissionsOrBuilder getUsersAndPermissionsOrBuilder() { - return usersAndPermissions_; - } - - private void initFields() { - name_ = ""; - table_ = ""; - creationTime_ = 0L; - type_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type.FLUSH; - version_ = 0; - owner_ = ""; - usersAndPermissions_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.getDefaultInstance(); - } - private byte memoizedIsInitialized = -1; - public final boolean isInitialized() { - byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; - - if (!hasName()) { - memoizedIsInitialized = 0; - return false; - } - if (hasUsersAndPermissions()) { - if (!getUsersAndPermissions().isInitialized()) { - memoizedIsInitialized = 0; - return false; - } - } - memoizedIsInitialized = 1; - return true; - } - - public void writeTo(com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getNameBytes()); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, getTableBytes()); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeInt64(3, creationTime_); - } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - output.writeEnum(4, type_.getNumber()); - } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - output.writeInt32(5, version_); - } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - output.writeBytes(6, getOwnerBytes()); - } - if (((bitField0_ & 0x00000040) == 0x00000040)) { - output.writeMessage(7, usersAndPermissions_); - } - getUnknownFields().writeTo(output); - } - - private int memoizedSerializedSize = -1; - public int getSerializedSize() { - int size = memoizedSerializedSize; - if (size != -1) return size; - - size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getNameBytes()); - } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, getTableBytes()); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream - .computeInt64Size(3, creationTime_); - } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - size += com.google.protobuf.CodedOutputStream - .computeEnumSize(4, type_.getNumber()); - } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - size += com.google.protobuf.CodedOutputStream - .computeInt32Size(5, version_); - } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(6, getOwnerBytes()); - } - if (((bitField0_ & 0x00000040) == 0x00000040)) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(7, usersAndPermissions_); - } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; - return size; - } - - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - - @java.lang.Override - public boolean equals(final java.lang.Object obj) { - if (obj == this) { - return true; - } - if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription)) { - return super.equals(obj); - } - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription other = (org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription) obj; - - boolean result = true; - result = result && (hasName() == other.hasName()); - if (hasName()) { - result = result && getName() - .equals(other.getName()); - } - result = result && (hasTable() == other.hasTable()); - if (hasTable()) { - result = result && getTable() - .equals(other.getTable()); - } - result = result && (hasCreationTime() == other.hasCreationTime()); - if (hasCreationTime()) { - result = result && (getCreationTime() - == other.getCreationTime()); - } - result = result && (hasType() == other.hasType()); - if (hasType()) { - result = result && - (getType() == other.getType()); - } - result = result && (hasVersion() == other.hasVersion()); - if (hasVersion()) { - result = result && (getVersion() - == other.getVersion()); - } - result = result && (hasOwner() == other.hasOwner()); - if (hasOwner()) { - result = result && getOwner() - .equals(other.getOwner()); - } - result = result && (hasUsersAndPermissions() == other.hasUsersAndPermissions()); - if (hasUsersAndPermissions()) { - result = result && getUsersAndPermissions() - .equals(other.getUsersAndPermissions()); - } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; - } - - private int memoizedHashCode = 0; - @java.lang.Override - public int hashCode() { - if (memoizedHashCode != 0) { - return memoizedHashCode; - } - int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasName()) { - hash = (37 * hash) + NAME_FIELD_NUMBER; - hash = (53 * hash) + getName().hashCode(); - } - if (hasTable()) { - hash = (37 * hash) + TABLE_FIELD_NUMBER; - hash = (53 * hash) + getTable().hashCode(); - } - if (hasCreationTime()) { - hash = (37 * hash) + CREATION_TIME_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getCreationTime()); - } - if (hasType()) { - hash = (37 * hash) + TYPE_FIELD_NUMBER; - hash = (53 * hash) + hashEnum(getType()); - } - if (hasVersion()) { - hash = (37 * hash) + VERSION_FIELD_NUMBER; - hash = (53 * hash) + getVersion(); - } - if (hasOwner()) { - hash = (37 * hash) + OWNER_FIELD_NUMBER; - hash = (53 * hash) + getOwner().hashCode(); - } - if (hasUsersAndPermissions()) { - hash = (37 * hash) + USERS_AND_PERMISSIONS_FIELD_NUMBER; - hash = (53 * hash) + getUsersAndPermissions().hashCode(); - } - hash = (29 * hash) + getUnknownFields().hashCode(); - memoizedHashCode = hash; - return hash; - } - - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseFrom( - byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseFrom(java.io.InputStream input) - throws java.io.IOException { - return PARSER.parseFrom(input); - } - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); - } - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseDelimitedFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); - } - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseFrom( - com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return PARSER.parseFrom(input); - } - public static org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parseFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); - } - - public static Builder newBuilder() { return Builder.create(); } - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription prototype) { - return newBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { return newBuilder(this); } - - @java.lang.Override - protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - Builder builder = new Builder(parent); - return builder; - } - /** - * Protobuf type {@code hbase.pb.SnapshotDescription} - * - *
-     **
-     * Description of the snapshot to take
-     * 
- */ - public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder { - public static final com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotDescription_descriptor; - } - - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotDescription_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.class, org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder.class); - } - - // Construct using org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.newBuilder() - private Builder() { - maybeForceBuilderInitialization(); - } - - private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { - super(parent); - maybeForceBuilderInitialization(); - } - private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { - getUsersAndPermissionsFieldBuilder(); - } - } - private static Builder create() { - return new Builder(); - } - - public Builder clear() { - super.clear(); - name_ = ""; - bitField0_ = (bitField0_ & ~0x00000001); - table_ = ""; - bitField0_ = (bitField0_ & ~0x00000002); - creationTime_ = 0L; - bitField0_ = (bitField0_ & ~0x00000004); - type_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type.FLUSH; - bitField0_ = (bitField0_ & ~0x00000008); - version_ = 0; - bitField0_ = (bitField0_ & ~0x00000010); - owner_ = ""; - bitField0_ = (bitField0_ & ~0x00000020); - if (usersAndPermissionsBuilder_ == null) { - usersAndPermissions_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.getDefaultInstance(); - } else { - usersAndPermissionsBuilder_.clear(); - } - bitField0_ = (bitField0_ & ~0x00000040); - return this; - } - - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - - public com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.internal_static_hbase_pb_SnapshotDescription_descriptor; - } - - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription getDefaultInstanceForType() { - return org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance(); - } - - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription build() { - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException(result); - } - return result; - } - - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription buildPartial() { - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription result = new org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription(this); - int from_bitField0_ = bitField0_; - int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { - to_bitField0_ |= 0x00000001; - } - result.name_ = name_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { - to_bitField0_ |= 0x00000002; - } - result.table_ = table_; - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { - to_bitField0_ |= 0x00000004; - } - result.creationTime_ = creationTime_; - if (((from_bitField0_ & 0x00000008) == 0x00000008)) { - to_bitField0_ |= 0x00000008; - } - result.type_ = type_; - if (((from_bitField0_ & 0x00000010) == 0x00000010)) { - to_bitField0_ |= 0x00000010; - } - result.version_ = version_; - if (((from_bitField0_ & 0x00000020) == 0x00000020)) { - to_bitField0_ |= 0x00000020; - } - result.owner_ = owner_; - if (((from_bitField0_ & 0x00000040) == 0x00000040)) { - to_bitField0_ |= 0x00000040; - } - if (usersAndPermissionsBuilder_ == null) { - result.usersAndPermissions_ = usersAndPermissions_; - } else { - result.usersAndPermissions_ = usersAndPermissionsBuilder_.build(); - } - result.bitField0_ = to_bitField0_; - onBuilt(); - return result; - } - - public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription) { - return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription other) { - if (other == org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance()) return this; - if (other.hasName()) { - bitField0_ |= 0x00000001; - name_ = other.name_; - onChanged(); - } - if (other.hasTable()) { - bitField0_ |= 0x00000002; - table_ = other.table_; - onChanged(); - } - if (other.hasCreationTime()) { - setCreationTime(other.getCreationTime()); - } - if (other.hasType()) { - setType(other.getType()); - } - if (other.hasVersion()) { - setVersion(other.getVersion()); - } - if (other.hasOwner()) { - bitField0_ |= 0x00000020; - owner_ = other.owner_; - onChanged(); - } - if (other.hasUsersAndPermissions()) { - mergeUsersAndPermissions(other.getUsersAndPermissions()); - } - this.mergeUnknownFields(other.getUnknownFields()); - return this; - } - - public final boolean isInitialized() { - if (!hasName()) { - - return false; - } - if (hasUsersAndPermissions()) { - if (!getUsersAndPermissions().isInitialized()) { - - return false; - } - } - return true; - } - - public Builder mergeFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription parsedMessage = null; - try { - parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription) e.getUnfinishedMessage(); - throw e; - } finally { - if (parsedMessage != null) { - mergeFrom(parsedMessage); - } - } - return this; - } - private int bitField0_; - - // required string name = 1; - private java.lang.Object name_ = ""; - /** - * required string name = 1; - */ - public boolean hasName() { - return ((bitField0_ & 0x00000001) == 0x00000001); - } - /** - * required string name = 1; - */ - public java.lang.String getName() { - java.lang.Object ref = name_; - if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - name_ = s; - return s; - } else { - return (java.lang.String) ref; - } - } - /** - * required string name = 1; - */ - public com.google.protobuf.ByteString - getNameBytes() { - java.lang.Object ref = name_; - if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - name_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - /** - * required string name = 1; - */ - public Builder setName( - java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000001; - name_ = value; - onChanged(); - return this; - } - /** - * required string name = 1; - */ - public Builder clearName() { - bitField0_ = (bitField0_ & ~0x00000001); - name_ = getDefaultInstance().getName(); - onChanged(); - return this; - } - /** - * required string name = 1; - */ - public Builder setNameBytes( - com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000001; - name_ = value; - onChanged(); - return this; - } - - // optional string table = 2; - private java.lang.Object table_ = ""; - /** - * optional string table = 2; - * - *
-       * not needed for delete, but checked for in taking snapshot
-       * 
- */ - public boolean hasTable() { - return ((bitField0_ & 0x00000002) == 0x00000002); - } - /** - * optional string table = 2; - * - *
-       * not needed for delete, but checked for in taking snapshot
-       * 
- */ - public java.lang.String getTable() { - java.lang.Object ref = table_; - if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - table_ = s; - return s; - } else { - return (java.lang.String) ref; - } - } - /** - * optional string table = 2; - * - *
-       * not needed for delete, but checked for in taking snapshot
-       * 
- */ - public com.google.protobuf.ByteString - getTableBytes() { - java.lang.Object ref = table_; - if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - table_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - /** - * optional string table = 2; - * - *
-       * not needed for delete, but checked for in taking snapshot
-       * 
- */ - public Builder setTable( - java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000002; - table_ = value; - onChanged(); - return this; - } - /** - * optional string table = 2; - * - *
-       * not needed for delete, but checked for in taking snapshot
-       * 
- */ - public Builder clearTable() { - bitField0_ = (bitField0_ & ~0x00000002); - table_ = getDefaultInstance().getTable(); - onChanged(); - return this; - } - /** - * optional string table = 2; - * - *
-       * not needed for delete, but checked for in taking snapshot
-       * 
- */ - public Builder setTableBytes( - com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000002; - table_ = value; - onChanged(); - return this; - } - - // optional int64 creation_time = 3 [default = 0]; - private long creationTime_ ; - /** - * optional int64 creation_time = 3 [default = 0]; - */ - public boolean hasCreationTime() { - return ((bitField0_ & 0x00000004) == 0x00000004); - } - /** - * optional int64 creation_time = 3 [default = 0]; - */ - public long getCreationTime() { - return creationTime_; - } - /** - * optional int64 creation_time = 3 [default = 0]; - */ - public Builder setCreationTime(long value) { - bitField0_ |= 0x00000004; - creationTime_ = value; - onChanged(); - return this; - } - /** - * optional int64 creation_time = 3 [default = 0]; - */ - public Builder clearCreationTime() { - bitField0_ = (bitField0_ & ~0x00000004); - creationTime_ = 0L; - onChanged(); - return this; - } - - // optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - private org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type type_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type.FLUSH; - /** - * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - */ - public boolean hasType() { - return ((bitField0_ & 0x00000008) == 0x00000008); - } - /** - * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - */ - public org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type getType() { - return type_; - } - /** - * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - */ - public Builder setType(org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000008; - type_ = value; - onChanged(); - return this; - } - /** - * optional .hbase.pb.SnapshotDescription.Type type = 4 [default = FLUSH]; - */ - public Builder clearType() { - bitField0_ = (bitField0_ & ~0x00000008); - type_ = org.apache.hadoop.hbase.protobuf.generated.SnapshotProtos.SnapshotDescription.Type.FLUSH; - onChanged(); - return this; - } - - // optional int32 version = 5; - private int version_ ; - /** - * optional int32 version = 5; - */ - public boolean hasVersion() { - return ((bitField0_ & 0x00000010) == 0x00000010); - } - /** - * optional int32 version = 5; - */ - public int getVersion() { - return version_; - } - /** - * optional int32 version = 5; - */ - public Builder setVersion(int value) { - bitField0_ |= 0x00000010; - version_ = value; - onChanged(); - return this; - } - /** - * optional int32 version = 5; - */ - public Builder clearVersion() { - bitField0_ = (bitField0_ & ~0x00000010); - version_ = 0; - onChanged(); - return this; - } - - // optional string owner = 6; - private java.lang.Object owner_ = ""; - /** - * optional string owner = 6; - */ - public boolean hasOwner() { - return ((bitField0_ & 0x00000020) == 0x00000020); - } - /** - * optional string owner = 6; - */ - public java.lang.String getOwner() { - java.lang.Object ref = owner_; - if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - owner_ = s; - return s; - } else { - return (java.lang.String) ref; - } - } - /** - * optional string owner = 6; - */ - public com.google.protobuf.ByteString - getOwnerBytes() { - java.lang.Object ref = owner_; - if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - owner_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - /** - * optional string owner = 6; - */ - public Builder setOwner( - java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000020; - owner_ = value; - onChanged(); - return this; - } - /** - * optional string owner = 6; - */ - public Builder clearOwner() { - bitField0_ = (bitField0_ & ~0x00000020); - owner_ = getDefaultInstance().getOwner(); - onChanged(); - return this; - } - /** - * optional string owner = 6; - */ - public Builder setOwnerBytes( - com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000020; - owner_ = value; - onChanged(); - return this; - } - - // optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - private org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions usersAndPermissions_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.Builder, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissionsOrBuilder> usersAndPermissionsBuilder_; - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public boolean hasUsersAndPermissions() { - return ((bitField0_ & 0x00000040) == 0x00000040); - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions getUsersAndPermissions() { - if (usersAndPermissionsBuilder_ == null) { - return usersAndPermissions_; - } else { - return usersAndPermissionsBuilder_.getMessage(); - } - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public Builder setUsersAndPermissions(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions value) { - if (usersAndPermissionsBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - usersAndPermissions_ = value; - onChanged(); - } else { - usersAndPermissionsBuilder_.setMessage(value); - } - bitField0_ |= 0x00000040; - return this; - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public Builder setUsersAndPermissions( - org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.Builder builderForValue) { - if (usersAndPermissionsBuilder_ == null) { - usersAndPermissions_ = builderForValue.build(); - onChanged(); - } else { - usersAndPermissionsBuilder_.setMessage(builderForValue.build()); - } - bitField0_ |= 0x00000040; - return this; - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public Builder mergeUsersAndPermissions(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions value) { - if (usersAndPermissionsBuilder_ == null) { - if (((bitField0_ & 0x00000040) == 0x00000040) && - usersAndPermissions_ != org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.getDefaultInstance()) { - usersAndPermissions_ = - org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.newBuilder(usersAndPermissions_).mergeFrom(value).buildPartial(); - } else { - usersAndPermissions_ = value; - } - onChanged(); - } else { - usersAndPermissionsBuilder_.mergeFrom(value); - } - bitField0_ |= 0x00000040; - return this; - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public Builder clearUsersAndPermissions() { - if (usersAndPermissionsBuilder_ == null) { - usersAndPermissions_ = org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.getDefaultInstance(); - onChanged(); - } else { - usersAndPermissionsBuilder_.clear(); - } - bitField0_ = (bitField0_ & ~0x00000040); - return this; - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.Builder getUsersAndPermissionsBuilder() { - bitField0_ |= 0x00000040; - onChanged(); - return getUsersAndPermissionsFieldBuilder().getBuilder(); - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - public org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissionsOrBuilder getUsersAndPermissionsOrBuilder() { - if (usersAndPermissionsBuilder_ != null) { - return usersAndPermissionsBuilder_.getMessageOrBuilder(); - } else { - return usersAndPermissions_; - } - } - /** - * optional .hbase.pb.UsersAndPermissions users_and_permissions = 7; - */ - private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.Builder, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissionsOrBuilder> - getUsersAndPermissionsFieldBuilder() { - if (usersAndPermissionsBuilder_ == null) { - usersAndPermissionsBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions.Builder, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissionsOrBuilder>( - usersAndPermissions_, - getParentForChildren(), - isClean()); - usersAndPermissions_ = null; - } - return usersAndPermissionsBuilder_; - } - - // @@protoc_insertion_point(builder_scope:hbase.pb.SnapshotDescription) - } - - static { - defaultInstance = new SnapshotDescription(true); - defaultInstance.initFields(); - } - - // @@protoc_insertion_point(class_scope:hbase.pb.SnapshotDescription) - } - public interface SnapshotFileInfoOrBuilder extends com.google.protobuf.MessageOrBuilder { @@ -6174,11 +4686,6 @@ public final class SnapshotProtos { } private static com.google.protobuf.Descriptors.Descriptor - internal_static_hbase_pb_SnapshotDescription_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable - internal_static_hbase_pb_SnapshotDescription_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SnapshotFileInfo_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable @@ -6212,52 +4719,39 @@ public final class SnapshotProtos { descriptor; static { java.lang.String[] descriptorData = { - "\n\016Snapshot.proto\022\010hbase.pb\032\023AccessContro" + - "l.proto\032\010FS.proto\032\013HBase.proto\"\223\002\n\023Snaps" + - "hotDescription\022\014\n\004name\030\001 \002(\t\022\r\n\005table\030\002 " + - "\001(\t\022\030\n\rcreation_time\030\003 \001(\003:\0010\0227\n\004type\030\004 " + - "\001(\0162\".hbase.pb.SnapshotDescription.Type:" + - "\005FLUSH\022\017\n\007version\030\005 \001(\005\022\r\n\005owner\030\006 \001(\t\022<" + - "\n\025users_and_permissions\030\007 \001(\0132\035.hbase.pb" + - ".UsersAndPermissions\".\n\004Type\022\014\n\010DISABLED" + - "\020\000\022\t\n\005FLUSH\020\001\022\r\n\tSKIPFLUSH\020\002\"\222\001\n\020Snapsho" + - "tFileInfo\022-\n\004type\030\001 \002(\0162\037.hbase.pb.Snaps", - "hotFileInfo.Type\022\r\n\005hfile\030\003 \001(\t\022\022\n\nwal_s" + - "erver\030\004 \001(\t\022\020\n\010wal_name\030\005 \001(\t\"\032\n\004Type\022\t\n" + - "\005HFILE\020\001\022\007\n\003WAL\020\002\"\323\002\n\026SnapshotRegionMani" + - "fest\022\017\n\007version\030\001 \001(\005\022)\n\013region_info\030\002 \002" + - "(\0132\024.hbase.pb.RegionInfo\022B\n\014family_files" + - "\030\003 \003(\0132,.hbase.pb.SnapshotRegionManifest" + - ".FamilyFiles\032T\n\tStoreFile\022\014\n\004name\030\001 \002(\t\022" + - "&\n\treference\030\002 \001(\0132\023.hbase.pb.Reference\022" + - "\021\n\tfile_size\030\003 \001(\004\032c\n\013FamilyFiles\022\023\n\013fam" + - "ily_name\030\001 \002(\014\022?\n\013store_files\030\002 \003(\0132*.hb", - "ase.pb.SnapshotRegionManifest.StoreFile\"" + - "\177\n\024SnapshotDataManifest\022+\n\014table_schema\030" + - "\001 \002(\0132\025.hbase.pb.TableSchema\022:\n\020region_m" + - "anifests\030\002 \003(\0132 .hbase.pb.SnapshotRegion" + - "ManifestBD\n*org.apache.hadoop.hbase.prot" + - "obuf.generatedB\016SnapshotProtosH\001\210\001\001\240\001\001" + "\n\016Snapshot.proto\022\010hbase.pb\032\010FS.proto\032\013HB" + + "ase.proto\"\222\001\n\020SnapshotFileInfo\022-\n\004type\030\001" + + " \002(\0162\037.hbase.pb.SnapshotFileInfo.Type\022\r\n" + + "\005hfile\030\003 \001(\t\022\022\n\nwal_server\030\004 \001(\t\022\020\n\010wal_" + + "name\030\005 \001(\t\"\032\n\004Type\022\t\n\005HFILE\020\001\022\007\n\003WAL\020\002\"\323" + + "\002\n\026SnapshotRegionManifest\022\017\n\007version\030\001 \001" + + "(\005\022)\n\013region_info\030\002 \002(\0132\024.hbase.pb.Regio" + + "nInfo\022B\n\014family_files\030\003 \003(\0132,.hbase.pb.S" + + "napshotRegionManifest.FamilyFiles\032T\n\tSto" + + "reFile\022\014\n\004name\030\001 \002(\t\022&\n\treference\030\002 \001(\0132", + "\023.hbase.pb.Reference\022\021\n\tfile_size\030\003 \001(\004\032" + + "c\n\013FamilyFiles\022\023\n\013family_name\030\001 \002(\014\022?\n\013s" + + "tore_files\030\002 \003(\0132*.hbase.pb.SnapshotRegi" + + "onManifest.StoreFile\"\177\n\024SnapshotDataMani" + + "fest\022+\n\014table_schema\030\001 \002(\0132\025.hbase.pb.Ta" + + "bleSchema\022:\n\020region_manifests\030\002 \003(\0132 .hb" + + "ase.pb.SnapshotRegionManifestBD\n*org.apa" + + "che.hadoop.hbase.protobuf.generatedB\016Sna" + + "pshotProtosH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { public com.google.protobuf.ExtensionRegistry assignDescriptors( com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; - internal_static_hbase_pb_SnapshotDescription_descriptor = - getDescriptor().getMessageTypes().get(0); - internal_static_hbase_pb_SnapshotDescription_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_hbase_pb_SnapshotDescription_descriptor, - new java.lang.String[] { "Name", "Table", "CreationTime", "Type", "Version", "Owner", "UsersAndPermissions", }); internal_static_hbase_pb_SnapshotFileInfo_descriptor = - getDescriptor().getMessageTypes().get(1); + getDescriptor().getMessageTypes().get(0); internal_static_hbase_pb_SnapshotFileInfo_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_SnapshotFileInfo_descriptor, new java.lang.String[] { "Type", "Hfile", "WalServer", "WalName", }); internal_static_hbase_pb_SnapshotRegionManifest_descriptor = - getDescriptor().getMessageTypes().get(2); + getDescriptor().getMessageTypes().get(1); internal_static_hbase_pb_SnapshotRegionManifest_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_SnapshotRegionManifest_descriptor, @@ -6275,7 +4769,7 @@ public final class SnapshotProtos { internal_static_hbase_pb_SnapshotRegionManifest_FamilyFiles_descriptor, new java.lang.String[] { "FamilyName", "StoreFiles", }); internal_static_hbase_pb_SnapshotDataManifest_descriptor = - getDescriptor().getMessageTypes().get(3); + getDescriptor().getMessageTypes().get(2); internal_static_hbase_pb_SnapshotDataManifest_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_hbase_pb_SnapshotDataManifest_descriptor, @@ -6286,7 +4780,6 @@ public final class SnapshotProtos { com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { - org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.FSProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(), }, assigner);