Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id 1A114200AEE for ; Wed, 4 May 2016 00:36:57 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id 18ABE1609F5; Wed, 4 May 2016 00:36:57 +0200 (CEST) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id C550A1609F7 for ; Wed, 4 May 2016 00:36:54 +0200 (CEST) Received: (qmail 64309 invoked by uid 500); 3 May 2016 22:36:53 -0000 Mailing-List: contact commits-help@hive.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: hive-dev@hive.apache.org Delivered-To: mailing list commits@hive.apache.org Received: (qmail 64098 invoked by uid 99); 3 May 2016 22:36:53 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 03 May 2016 22:36:53 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id 8D60EDFDEC; Tue, 3 May 2016 22:36:53 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: sershe@apache.org To: commits@hive.apache.org Date: Tue, 03 May 2016 22:36:54 -0000 Message-Id: <7d9fbe0e46ca4f16be309a45e77e0836@git.apache.org> In-Reply-To: <2f1638b1ee1747fc9cce184b890d4b11@git.apache.org> References: <2f1638b1ee1747fc9cce184b890d4b11@git.apache.org> X-Mailer: ASF-Git Admin Mailer Subject: [2/3] hive git commit: HIVE-13442 : LLAP: refactor submit API to be amenable to signing (Sergey Shelukhin, reviewed by Siddharth Seth) archived-at: Tue, 03 May 2016 22:36:57 -0000 http://git-wip-us.apache.org/repos/asf/hive/blob/0b5c27fd/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java ---------------------------------------------------------------------- diff --git a/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java b/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java index 820f6be..16531e3 100644 --- a/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java +++ b/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java @@ -3212,202 +3212,73 @@ public final class LlapDaemonProtocolProtos { // @@protoc_insertion_point(class_scope:GroupInputSpecProto) } - public interface FragmentSpecProtoOrBuilder + public interface VertexIdentifierOrBuilder extends com.google.protobuf.MessageOrBuilder { - // optional string fragment_identifier_string = 1; + // optional string application_id_string = 1; /** - * optional string fragment_identifier_string = 1; + * optional string application_id_string = 1; */ - boolean hasFragmentIdentifierString(); + boolean hasApplicationIdString(); /** - * optional string fragment_identifier_string = 1; + * optional string application_id_string = 1; */ - java.lang.String getFragmentIdentifierString(); + java.lang.String getApplicationIdString(); /** - * optional string fragment_identifier_string = 1; + * optional string application_id_string = 1; */ com.google.protobuf.ByteString - getFragmentIdentifierStringBytes(); + getApplicationIdStringBytes(); - // optional string dag_name = 2; - /** - * optional string dag_name = 2; - */ - boolean hasDagName(); + // optional int32 app_attempt_number = 2; /** - * optional string dag_name = 2; + * optional int32 app_attempt_number = 2; */ - java.lang.String getDagName(); + boolean hasAppAttemptNumber(); /** - * optional string dag_name = 2; + * optional int32 app_attempt_number = 2; */ - com.google.protobuf.ByteString - getDagNameBytes(); + int getAppAttemptNumber(); - // optional int32 dag_id = 11; + // optional int32 dag_id = 3; /** - * optional int32 dag_id = 11; + * optional int32 dag_id = 3; */ boolean hasDagId(); /** - * optional int32 dag_id = 11; + * optional int32 dag_id = 3; */ int getDagId(); - // optional string vertex_name = 3; - /** - * optional string vertex_name = 3; - */ - boolean hasVertexName(); - /** - * optional string vertex_name = 3; - */ - java.lang.String getVertexName(); - /** - * optional string vertex_name = 3; - */ - com.google.protobuf.ByteString - getVertexNameBytes(); - - // optional .EntityDescriptorProto processor_descriptor = 4; - /** - * optional .EntityDescriptorProto processor_descriptor = 4; - */ - boolean hasProcessorDescriptor(); - /** - * optional .EntityDescriptorProto processor_descriptor = 4; - */ - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getProcessorDescriptor(); - /** - * optional .EntityDescriptorProto processor_descriptor = 4; - */ - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getProcessorDescriptorOrBuilder(); - - // repeated .IOSpecProto input_specs = 5; - /** - * repeated .IOSpecProto input_specs = 5; - */ - java.util.List - getInputSpecsList(); - /** - * repeated .IOSpecProto input_specs = 5; - */ - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getInputSpecs(int index); - /** - * repeated .IOSpecProto input_specs = 5; - */ - int getInputSpecsCount(); - /** - * repeated .IOSpecProto input_specs = 5; - */ - java.util.List - getInputSpecsOrBuilderList(); - /** - * repeated .IOSpecProto input_specs = 5; - */ - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getInputSpecsOrBuilder( - int index); - - // repeated .IOSpecProto output_specs = 6; - /** - * repeated .IOSpecProto output_specs = 6; - */ - java.util.List - getOutputSpecsList(); - /** - * repeated .IOSpecProto output_specs = 6; - */ - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getOutputSpecs(int index); - /** - * repeated .IOSpecProto output_specs = 6; - */ - int getOutputSpecsCount(); - /** - * repeated .IOSpecProto output_specs = 6; - */ - java.util.List - getOutputSpecsOrBuilderList(); - /** - * repeated .IOSpecProto output_specs = 6; - */ - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getOutputSpecsOrBuilder( - int index); - - // repeated .GroupInputSpecProto grouped_input_specs = 7; - /** - * repeated .GroupInputSpecProto grouped_input_specs = 7; - */ - java.util.List - getGroupedInputSpecsList(); - /** - * repeated .GroupInputSpecProto grouped_input_specs = 7; - */ - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getGroupedInputSpecs(int index); - /** - * repeated .GroupInputSpecProto grouped_input_specs = 7; - */ - int getGroupedInputSpecsCount(); - /** - * repeated .GroupInputSpecProto grouped_input_specs = 7; - */ - java.util.List - getGroupedInputSpecsOrBuilderList(); - /** - * repeated .GroupInputSpecProto grouped_input_specs = 7; - */ - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder getGroupedInputSpecsOrBuilder( - int index); - - // optional int32 vertex_parallelism = 8; - /** - * optional int32 vertex_parallelism = 8; - */ - boolean hasVertexParallelism(); - /** - * optional int32 vertex_parallelism = 8; - */ - int getVertexParallelism(); - - // optional int32 fragment_number = 9; - /** - * optional int32 fragment_number = 9; - */ - boolean hasFragmentNumber(); - /** - * optional int32 fragment_number = 9; - */ - int getFragmentNumber(); - - // optional int32 attempt_number = 10; + // optional int32 vertex_id = 4; /** - * optional int32 attempt_number = 10; + * optional int32 vertex_id = 4; */ - boolean hasAttemptNumber(); + boolean hasVertexId(); /** - * optional int32 attempt_number = 10; + * optional int32 vertex_id = 4; */ - int getAttemptNumber(); + int getVertexId(); } /** - * Protobuf type {@code FragmentSpecProto} + * Protobuf type {@code VertexIdentifier} */ - public static final class FragmentSpecProto extends + public static final class VertexIdentifier extends com.google.protobuf.GeneratedMessage - implements FragmentSpecProtoOrBuilder { - // Use FragmentSpecProto.newBuilder() to construct. - private FragmentSpecProto(com.google.protobuf.GeneratedMessage.Builder builder) { + implements VertexIdentifierOrBuilder { + // Use VertexIdentifier.newBuilder() to construct. + private VertexIdentifier(com.google.protobuf.GeneratedMessage.Builder builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } - private FragmentSpecProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + private VertexIdentifier(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - private static final FragmentSpecProto defaultInstance; - public static FragmentSpecProto getDefaultInstance() { + private static final VertexIdentifier defaultInstance; + public static VertexIdentifier getDefaultInstance() { return defaultInstance; } - public FragmentSpecProto getDefaultInstanceForType() { + public VertexIdentifier getDefaultInstanceForType() { return defaultInstance; } @@ -3417,7 +3288,7 @@ public final class LlapDaemonProtocolProtos { getUnknownFields() { return this.unknownFields; } - private FragmentSpecProto( + private VertexIdentifier( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { @@ -3442,76 +3313,24 @@ public final class LlapDaemonProtocolProtos { } case 10: { bitField0_ |= 0x00000001; - fragmentIdentifierString_ = input.readBytes(); + applicationIdString_ = input.readBytes(); break; } - case 18: { + case 16: { bitField0_ |= 0x00000002; - dagName_ = input.readBytes(); - break; - } - case 26: { - bitField0_ |= 0x00000008; - vertexName_ = input.readBytes(); - break; - } - case 34: { - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder subBuilder = null; - if (((bitField0_ & 0x00000010) == 0x00000010)) { - subBuilder = processorDescriptor_.toBuilder(); - } - processorDescriptor_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.PARSER, extensionRegistry); - if (subBuilder != null) { - subBuilder.mergeFrom(processorDescriptor_); - processorDescriptor_ = subBuilder.buildPartial(); - } - bitField0_ |= 0x00000010; - break; - } - case 42: { - if (!((mutable_bitField0_ & 0x00000020) == 0x00000020)) { - inputSpecs_ = new java.util.ArrayList(); - mutable_bitField0_ |= 0x00000020; - } - inputSpecs_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.PARSER, extensionRegistry)); - break; - } - case 50: { - if (!((mutable_bitField0_ & 0x00000040) == 0x00000040)) { - outputSpecs_ = new java.util.ArrayList(); - mutable_bitField0_ |= 0x00000040; - } - outputSpecs_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.PARSER, extensionRegistry)); - break; - } - case 58: { - if (!((mutable_bitField0_ & 0x00000080) == 0x00000080)) { - groupedInputSpecs_ = new java.util.ArrayList(); - mutable_bitField0_ |= 0x00000080; - } - groupedInputSpecs_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.PARSER, extensionRegistry)); - break; - } - case 64: { - bitField0_ |= 0x00000020; - vertexParallelism_ = input.readInt32(); - break; - } - case 72: { - bitField0_ |= 0x00000040; - fragmentNumber_ = input.readInt32(); - break; - } - case 80: { - bitField0_ |= 0x00000080; - attemptNumber_ = input.readInt32(); + appAttemptNumber_ = input.readInt32(); break; } - case 88: { + case 24: { bitField0_ |= 0x00000004; dagId_ = input.readInt32(); break; } + case 32: { + bitField0_ |= 0x00000008; + vertexId_ = input.readInt32(); + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { @@ -3520,61 +3339,52 @@ public final class LlapDaemonProtocolProtos { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { - if (((mutable_bitField0_ & 0x00000020) == 0x00000020)) { - inputSpecs_ = java.util.Collections.unmodifiableList(inputSpecs_); - } - if (((mutable_bitField0_ & 0x00000040) == 0x00000040)) { - outputSpecs_ = java.util.Collections.unmodifiableList(outputSpecs_); - } - if (((mutable_bitField0_ & 0x00000080) == 0x00000080)) { - groupedInputSpecs_ = java.util.Collections.unmodifiableList(groupedInputSpecs_); - } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentSpecProto_descriptor; + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexIdentifier_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentSpecProto_fieldAccessorTable + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexIdentifier_fieldAccessorTable .ensureFieldAccessorsInitialized( - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto.Builder.class); + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public FragmentSpecProto parsePartialFrom( + public static com.google.protobuf.Parser PARSER = + new com.google.protobuf.AbstractParser() { + public VertexIdentifier parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - return new FragmentSpecProto(input, extensionRegistry); + return new VertexIdentifier(input, extensionRegistry); } }; @java.lang.Override - public com.google.protobuf.Parser getParserForType() { + public com.google.protobuf.Parser getParserForType() { return PARSER; } private int bitField0_; - // optional string fragment_identifier_string = 1; - public static final int FRAGMENT_IDENTIFIER_STRING_FIELD_NUMBER = 1; - private java.lang.Object fragmentIdentifierString_; + // optional string application_id_string = 1; + public static final int APPLICATION_ID_STRING_FIELD_NUMBER = 1; + private java.lang.Object applicationIdString_; /** - * optional string fragment_identifier_string = 1; + * optional string application_id_string = 1; */ - public boolean hasFragmentIdentifierString() { + public boolean hasApplicationIdString() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** - * optional string fragment_identifier_string = 1; + * optional string application_id_string = 1; */ - public java.lang.String getFragmentIdentifierString() { - java.lang.Object ref = fragmentIdentifierString_; + public java.lang.String getApplicationIdString() { + java.lang.Object ref = applicationIdString_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { @@ -3582,320 +3392,81 @@ public final class LlapDaemonProtocolProtos { (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { - fragmentIdentifierString_ = s; + applicationIdString_ = s; } return s; } } /** - * optional string fragment_identifier_string = 1; + * optional string application_id_string = 1; */ public com.google.protobuf.ByteString - getFragmentIdentifierStringBytes() { - java.lang.Object ref = fragmentIdentifierString_; + getApplicationIdStringBytes() { + java.lang.Object ref = applicationIdString_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); - fragmentIdentifierString_ = b; + applicationIdString_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } - // optional string dag_name = 2; - public static final int DAG_NAME_FIELD_NUMBER = 2; - private java.lang.Object dagName_; + // optional int32 app_attempt_number = 2; + public static final int APP_ATTEMPT_NUMBER_FIELD_NUMBER = 2; + private int appAttemptNumber_; /** - * optional string dag_name = 2; + * optional int32 app_attempt_number = 2; */ - public boolean hasDagName() { + public boolean hasAppAttemptNumber() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** - * optional string dag_name = 2; - */ - public java.lang.String getDagName() { - java.lang.Object ref = dagName_; - if (ref instanceof java.lang.String) { - return (java.lang.String) ref; - } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; - java.lang.String s = bs.toStringUtf8(); - if (bs.isValidUtf8()) { - dagName_ = s; - } - return s; - } - } - /** - * optional string dag_name = 2; + * optional int32 app_attempt_number = 2; */ - public com.google.protobuf.ByteString - getDagNameBytes() { - java.lang.Object ref = dagName_; - if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - dagName_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } + public int getAppAttemptNumber() { + return appAttemptNumber_; } - // optional int32 dag_id = 11; - public static final int DAG_ID_FIELD_NUMBER = 11; + // optional int32 dag_id = 3; + public static final int DAG_ID_FIELD_NUMBER = 3; private int dagId_; /** - * optional int32 dag_id = 11; + * optional int32 dag_id = 3; */ public boolean hasDagId() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** - * optional int32 dag_id = 11; + * optional int32 dag_id = 3; */ public int getDagId() { return dagId_; } - // optional string vertex_name = 3; - public static final int VERTEX_NAME_FIELD_NUMBER = 3; - private java.lang.Object vertexName_; + // optional int32 vertex_id = 4; + public static final int VERTEX_ID_FIELD_NUMBER = 4; + private int vertexId_; /** - * optional string vertex_name = 3; + * optional int32 vertex_id = 4; */ - public boolean hasVertexName() { + public boolean hasVertexId() { return ((bitField0_ & 0x00000008) == 0x00000008); } /** - * optional string vertex_name = 3; - */ - public java.lang.String getVertexName() { - java.lang.Object ref = vertexName_; - if (ref instanceof java.lang.String) { - return (java.lang.String) ref; - } else { - com.google.protobuf.ByteString bs = - (com.google.protobuf.ByteString) ref; - java.lang.String s = bs.toStringUtf8(); - if (bs.isValidUtf8()) { - vertexName_ = s; - } - return s; - } - } - /** - * optional string vertex_name = 3; + * optional int32 vertex_id = 4; */ - public com.google.protobuf.ByteString - getVertexNameBytes() { - java.lang.Object ref = vertexName_; - if (ref instanceof java.lang.String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - vertexName_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - - // optional .EntityDescriptorProto processor_descriptor = 4; - public static final int PROCESSOR_DESCRIPTOR_FIELD_NUMBER = 4; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto processorDescriptor_; - /** - * optional .EntityDescriptorProto processor_descriptor = 4; - */ - public boolean hasProcessorDescriptor() { - return ((bitField0_ & 0x00000010) == 0x00000010); - } - /** - * optional .EntityDescriptorProto processor_descriptor = 4; - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getProcessorDescriptor() { - return processorDescriptor_; - } - /** - * optional .EntityDescriptorProto processor_descriptor = 4; - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getProcessorDescriptorOrBuilder() { - return processorDescriptor_; - } - - // repeated .IOSpecProto input_specs = 5; - public static final int INPUT_SPECS_FIELD_NUMBER = 5; - private java.util.List inputSpecs_; - /** - * repeated .IOSpecProto input_specs = 5; - */ - public java.util.List getInputSpecsList() { - return inputSpecs_; - } - /** - * repeated .IOSpecProto input_specs = 5; - */ - public java.util.List - getInputSpecsOrBuilderList() { - return inputSpecs_; - } - /** - * repeated .IOSpecProto input_specs = 5; - */ - public int getInputSpecsCount() { - return inputSpecs_.size(); - } - /** - * repeated .IOSpecProto input_specs = 5; - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getInputSpecs(int index) { - return inputSpecs_.get(index); - } - /** - * repeated .IOSpecProto input_specs = 5; - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getInputSpecsOrBuilder( - int index) { - return inputSpecs_.get(index); - } - - // repeated .IOSpecProto output_specs = 6; - public static final int OUTPUT_SPECS_FIELD_NUMBER = 6; - private java.util.List outputSpecs_; - /** - * repeated .IOSpecProto output_specs = 6; - */ - public java.util.List getOutputSpecsList() { - return outputSpecs_; - } - /** - * repeated .IOSpecProto output_specs = 6; - */ - public java.util.List - getOutputSpecsOrBuilderList() { - return outputSpecs_; - } - /** - * repeated .IOSpecProto output_specs = 6; - */ - public int getOutputSpecsCount() { - return outputSpecs_.size(); - } - /** - * repeated .IOSpecProto output_specs = 6; - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getOutputSpecs(int index) { - return outputSpecs_.get(index); - } - /** - * repeated .IOSpecProto output_specs = 6; - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getOutputSpecsOrBuilder( - int index) { - return outputSpecs_.get(index); - } - - // repeated .GroupInputSpecProto grouped_input_specs = 7; - public static final int GROUPED_INPUT_SPECS_FIELD_NUMBER = 7; - private java.util.List groupedInputSpecs_; - /** - * repeated .GroupInputSpecProto grouped_input_specs = 7; - */ - public java.util.List getGroupedInputSpecsList() { - return groupedInputSpecs_; - } - /** - * repeated .GroupInputSpecProto grouped_input_specs = 7; - */ - public java.util.List - getGroupedInputSpecsOrBuilderList() { - return groupedInputSpecs_; - } - /** - * repeated .GroupInputSpecProto grouped_input_specs = 7; - */ - public int getGroupedInputSpecsCount() { - return groupedInputSpecs_.size(); - } - /** - * repeated .GroupInputSpecProto grouped_input_specs = 7; - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getGroupedInputSpecs(int index) { - return groupedInputSpecs_.get(index); - } - /** - * repeated .GroupInputSpecProto grouped_input_specs = 7; - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder getGroupedInputSpecsOrBuilder( - int index) { - return groupedInputSpecs_.get(index); - } - - // optional int32 vertex_parallelism = 8; - public static final int VERTEX_PARALLELISM_FIELD_NUMBER = 8; - private int vertexParallelism_; - /** - * optional int32 vertex_parallelism = 8; - */ - public boolean hasVertexParallelism() { - return ((bitField0_ & 0x00000020) == 0x00000020); - } - /** - * optional int32 vertex_parallelism = 8; - */ - public int getVertexParallelism() { - return vertexParallelism_; - } - - // optional int32 fragment_number = 9; - public static final int FRAGMENT_NUMBER_FIELD_NUMBER = 9; - private int fragmentNumber_; - /** - * optional int32 fragment_number = 9; - */ - public boolean hasFragmentNumber() { - return ((bitField0_ & 0x00000040) == 0x00000040); - } - /** - * optional int32 fragment_number = 9; - */ - public int getFragmentNumber() { - return fragmentNumber_; - } - - // optional int32 attempt_number = 10; - public static final int ATTEMPT_NUMBER_FIELD_NUMBER = 10; - private int attemptNumber_; - /** - * optional int32 attempt_number = 10; - */ - public boolean hasAttemptNumber() { - return ((bitField0_ & 0x00000080) == 0x00000080); - } - /** - * optional int32 attempt_number = 10; - */ - public int getAttemptNumber() { - return attemptNumber_; + public int getVertexId() { + return vertexId_; } private void initFields() { - fragmentIdentifierString_ = ""; - dagName_ = ""; + applicationIdString_ = ""; + appAttemptNumber_ = 0; dagId_ = 0; - vertexName_ = ""; - processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); - inputSpecs_ = java.util.Collections.emptyList(); - outputSpecs_ = java.util.Collections.emptyList(); - groupedInputSpecs_ = java.util.Collections.emptyList(); - vertexParallelism_ = 0; - fragmentNumber_ = 0; - attemptNumber_ = 0; + vertexId_ = 0; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { @@ -3910,37 +3481,16 @@ public final class LlapDaemonProtocolProtos { throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getFragmentIdentifierStringBytes()); + output.writeBytes(1, getApplicationIdStringBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, getDagNameBytes()); - } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - output.writeBytes(3, getVertexNameBytes()); - } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - output.writeMessage(4, processorDescriptor_); - } - for (int i = 0; i < inputSpecs_.size(); i++) { - output.writeMessage(5, inputSpecs_.get(i)); - } - for (int i = 0; i < outputSpecs_.size(); i++) { - output.writeMessage(6, outputSpecs_.get(i)); - } - for (int i = 0; i < groupedInputSpecs_.size(); i++) { - output.writeMessage(7, groupedInputSpecs_.get(i)); - } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - output.writeInt32(8, vertexParallelism_); - } - if (((bitField0_ & 0x00000040) == 0x00000040)) { - output.writeInt32(9, fragmentNumber_); - } - if (((bitField0_ & 0x00000080) == 0x00000080)) { - output.writeInt32(10, attemptNumber_); + output.writeInt32(2, appAttemptNumber_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeInt32(11, dagId_); + output.writeInt32(3, dagId_); + } + if (((bitField0_ & 0x00000008) == 0x00000008)) { + output.writeInt32(4, vertexId_); } getUnknownFields().writeTo(output); } @@ -3953,47 +3503,19 @@ public final class LlapDaemonProtocolProtos { size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getFragmentIdentifierStringBytes()); + .computeBytesSize(1, getApplicationIdStringBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, getDagNameBytes()); - } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(3, getVertexNameBytes()); - } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(4, processorDescriptor_); + .computeInt32Size(2, appAttemptNumber_); } - for (int i = 0; i < inputSpecs_.size(); i++) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(5, inputSpecs_.get(i)); - } - for (int i = 0; i < outputSpecs_.size(); i++) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(6, outputSpecs_.get(i)); - } - for (int i = 0; i < groupedInputSpecs_.size(); i++) { - size += com.google.protobuf.CodedOutputStream - .computeMessageSize(7, groupedInputSpecs_.get(i)); - } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - size += com.google.protobuf.CodedOutputStream - .computeInt32Size(8, vertexParallelism_); - } - if (((bitField0_ & 0x00000040) == 0x00000040)) { - size += com.google.protobuf.CodedOutputStream - .computeInt32Size(9, fragmentNumber_); - } - if (((bitField0_ & 0x00000080) == 0x00000080)) { + if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream - .computeInt32Size(10, attemptNumber_); + .computeInt32Size(3, dagId_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000008) == 0x00000008)) { size += com.google.protobuf.CodedOutputStream - .computeInt32Size(11, dagId_); + .computeInt32Size(4, vertexId_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; @@ -4012,57 +3534,31 @@ public final class LlapDaemonProtocolProtos { if (obj == this) { return true; } - if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto)) { + if (!(obj instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier)) { return super.equals(obj); } - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto) obj; + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier) obj; boolean result = true; - result = result && (hasFragmentIdentifierString() == other.hasFragmentIdentifierString()); - if (hasFragmentIdentifierString()) { - result = result && getFragmentIdentifierString() - .equals(other.getFragmentIdentifierString()); + result = result && (hasApplicationIdString() == other.hasApplicationIdString()); + if (hasApplicationIdString()) { + result = result && getApplicationIdString() + .equals(other.getApplicationIdString()); } - result = result && (hasDagName() == other.hasDagName()); - if (hasDagName()) { - result = result && getDagName() - .equals(other.getDagName()); + result = result && (hasAppAttemptNumber() == other.hasAppAttemptNumber()); + if (hasAppAttemptNumber()) { + result = result && (getAppAttemptNumber() + == other.getAppAttemptNumber()); } result = result && (hasDagId() == other.hasDagId()); if (hasDagId()) { result = result && (getDagId() == other.getDagId()); } - result = result && (hasVertexName() == other.hasVertexName()); - if (hasVertexName()) { - result = result && getVertexName() - .equals(other.getVertexName()); - } - result = result && (hasProcessorDescriptor() == other.hasProcessorDescriptor()); - if (hasProcessorDescriptor()) { - result = result && getProcessorDescriptor() - .equals(other.getProcessorDescriptor()); - } - result = result && getInputSpecsList() - .equals(other.getInputSpecsList()); - result = result && getOutputSpecsList() - .equals(other.getOutputSpecsList()); - result = result && getGroupedInputSpecsList() - .equals(other.getGroupedInputSpecsList()); - result = result && (hasVertexParallelism() == other.hasVertexParallelism()); - if (hasVertexParallelism()) { - result = result && (getVertexParallelism() - == other.getVertexParallelism()); - } - result = result && (hasFragmentNumber() == other.hasFragmentNumber()); - if (hasFragmentNumber()) { - result = result && (getFragmentNumber() - == other.getFragmentNumber()); - } - result = result && (hasAttemptNumber() == other.hasAttemptNumber()); - if (hasAttemptNumber()) { - result = result && (getAttemptNumber() - == other.getAttemptNumber()); + result = result && (hasVertexId() == other.hasVertexId()); + if (hasVertexId()) { + result = result && (getVertexId() + == other.getVertexId()); } result = result && getUnknownFields().equals(other.getUnknownFields()); @@ -4077,102 +3573,74 @@ public final class LlapDaemonProtocolProtos { } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); - if (hasFragmentIdentifierString()) { - hash = (37 * hash) + FRAGMENT_IDENTIFIER_STRING_FIELD_NUMBER; - hash = (53 * hash) + getFragmentIdentifierString().hashCode(); + if (hasApplicationIdString()) { + hash = (37 * hash) + APPLICATION_ID_STRING_FIELD_NUMBER; + hash = (53 * hash) + getApplicationIdString().hashCode(); } - if (hasDagName()) { - hash = (37 * hash) + DAG_NAME_FIELD_NUMBER; - hash = (53 * hash) + getDagName().hashCode(); + if (hasAppAttemptNumber()) { + hash = (37 * hash) + APP_ATTEMPT_NUMBER_FIELD_NUMBER; + hash = (53 * hash) + getAppAttemptNumber(); } if (hasDagId()) { hash = (37 * hash) + DAG_ID_FIELD_NUMBER; hash = (53 * hash) + getDagId(); } - if (hasVertexName()) { - hash = (37 * hash) + VERTEX_NAME_FIELD_NUMBER; - hash = (53 * hash) + getVertexName().hashCode(); - } - if (hasProcessorDescriptor()) { - hash = (37 * hash) + PROCESSOR_DESCRIPTOR_FIELD_NUMBER; - hash = (53 * hash) + getProcessorDescriptor().hashCode(); - } - if (getInputSpecsCount() > 0) { - hash = (37 * hash) + INPUT_SPECS_FIELD_NUMBER; - hash = (53 * hash) + getInputSpecsList().hashCode(); - } - if (getOutputSpecsCount() > 0) { - hash = (37 * hash) + OUTPUT_SPECS_FIELD_NUMBER; - hash = (53 * hash) + getOutputSpecsList().hashCode(); - } - if (getGroupedInputSpecsCount() > 0) { - hash = (37 * hash) + GROUPED_INPUT_SPECS_FIELD_NUMBER; - hash = (53 * hash) + getGroupedInputSpecsList().hashCode(); - } - if (hasVertexParallelism()) { - hash = (37 * hash) + VERTEX_PARALLELISM_FIELD_NUMBER; - hash = (53 * hash) + getVertexParallelism(); - } - if (hasFragmentNumber()) { - hash = (37 * hash) + FRAGMENT_NUMBER_FIELD_NUMBER; - hash = (53 * hash) + getFragmentNumber(); - } - if (hasAttemptNumber()) { - hash = (37 * hash) + ATTEMPT_NUMBER_FIELD_NUMBER; - hash = (53 * hash) + getAttemptNumber(); + if (hasVertexId()) { + hash = (37 * hash) + VERTEX_ID_FIELD_NUMBER; + hash = (53 * hash) + getVertexId(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto parseFrom( + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto parseFrom( + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto parseFrom(byte[] data) + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto parseFrom( + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto parseFrom(java.io.InputStream input) + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto parseFrom( + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto parseDelimitedFrom(java.io.InputStream input) + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto parseDelimitedFrom( + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto parseFrom( + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } - public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto parseFrom( + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -4181,7 +3649,7 @@ public final class LlapDaemonProtocolProtos { public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto prototype) { + public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @@ -4193,24 +3661,24 @@ public final class LlapDaemonProtocolProtos { return builder; } /** - * Protobuf type {@code FragmentSpecProto} + * Protobuf type {@code VertexIdentifier} */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProtoOrBuilder { + implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifierOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentSpecProto_descriptor; + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexIdentifier_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentSpecProto_fieldAccessorTable + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexIdentifier_fieldAccessorTable .ensureFieldAccessorsInitialized( - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto.Builder.class); + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier.Builder.class); } - // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto.newBuilder() + // Construct using org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier.newBuilder() private Builder() { maybeForceBuilderInitialization(); } @@ -4222,10 +3690,6 @@ public final class LlapDaemonProtocolProtos { } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { - getProcessorDescriptorFieldBuilder(); - getInputSpecsFieldBuilder(); - getOutputSpecsFieldBuilder(); - getGroupedInputSpecsFieldBuilder(); } } private static Builder create() { @@ -4234,44 +3698,14 @@ public final class LlapDaemonProtocolProtos { public Builder clear() { super.clear(); - fragmentIdentifierString_ = ""; + applicationIdString_ = ""; bitField0_ = (bitField0_ & ~0x00000001); - dagName_ = ""; + appAttemptNumber_ = 0; bitField0_ = (bitField0_ & ~0x00000002); dagId_ = 0; bitField0_ = (bitField0_ & ~0x00000004); - vertexName_ = ""; + vertexId_ = 0; bitField0_ = (bitField0_ & ~0x00000008); - if (processorDescriptorBuilder_ == null) { - processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); - } else { - processorDescriptorBuilder_.clear(); - } - bitField0_ = (bitField0_ & ~0x00000010); - if (inputSpecsBuilder_ == null) { - inputSpecs_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000020); - } else { - inputSpecsBuilder_.clear(); - } - if (outputSpecsBuilder_ == null) { - outputSpecs_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000040); - } else { - outputSpecsBuilder_.clear(); - } - if (groupedInputSpecsBuilder_ == null) { - groupedInputSpecs_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000080); - } else { - groupedInputSpecsBuilder_.clear(); - } - vertexParallelism_ = 0; - bitField0_ = (bitField0_ & ~0x00000100); - fragmentNumber_ = 0; - bitField0_ = (bitField0_ & ~0x00000200); - attemptNumber_ = 0; - bitField0_ = (bitField0_ & ~0x00000400); return this; } @@ -4281,33 +3715,33 @@ public final class LlapDaemonProtocolProtos { public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentSpecProto_descriptor; + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexIdentifier_descriptor; } - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto getDefaultInstanceForType() { - return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto.getDefaultInstance(); + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier getDefaultInstanceForType() { + return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier.getDefaultInstance(); } - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto build() { - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto result = buildPartial(); + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier build() { + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto buildPartial() { - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto(this); + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier buildPartial() { + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } - result.fragmentIdentifierString_ = fragmentIdentifierString_; + result.applicationIdString_ = applicationIdString_; if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } - result.dagName_ = dagName_; + result.appAttemptNumber_ = appAttemptNumber_; if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000004; } @@ -4315,177 +3749,36 @@ public final class LlapDaemonProtocolProtos { if (((from_bitField0_ & 0x00000008) == 0x00000008)) { to_bitField0_ |= 0x00000008; } - result.vertexName_ = vertexName_; - if (((from_bitField0_ & 0x00000010) == 0x00000010)) { - to_bitField0_ |= 0x00000010; - } - if (processorDescriptorBuilder_ == null) { - result.processorDescriptor_ = processorDescriptor_; - } else { - result.processorDescriptor_ = processorDescriptorBuilder_.build(); - } - if (inputSpecsBuilder_ == null) { - if (((bitField0_ & 0x00000020) == 0x00000020)) { - inputSpecs_ = java.util.Collections.unmodifiableList(inputSpecs_); - bitField0_ = (bitField0_ & ~0x00000020); - } - result.inputSpecs_ = inputSpecs_; - } else { - result.inputSpecs_ = inputSpecsBuilder_.build(); - } - if (outputSpecsBuilder_ == null) { - if (((bitField0_ & 0x00000040) == 0x00000040)) { - outputSpecs_ = java.util.Collections.unmodifiableList(outputSpecs_); - bitField0_ = (bitField0_ & ~0x00000040); - } - result.outputSpecs_ = outputSpecs_; - } else { - result.outputSpecs_ = outputSpecsBuilder_.build(); - } - if (groupedInputSpecsBuilder_ == null) { - if (((bitField0_ & 0x00000080) == 0x00000080)) { - groupedInputSpecs_ = java.util.Collections.unmodifiableList(groupedInputSpecs_); - bitField0_ = (bitField0_ & ~0x00000080); - } - result.groupedInputSpecs_ = groupedInputSpecs_; - } else { - result.groupedInputSpecs_ = groupedInputSpecsBuilder_.build(); - } - if (((from_bitField0_ & 0x00000100) == 0x00000100)) { - to_bitField0_ |= 0x00000020; - } - result.vertexParallelism_ = vertexParallelism_; - if (((from_bitField0_ & 0x00000200) == 0x00000200)) { - to_bitField0_ |= 0x00000040; - } - result.fragmentNumber_ = fragmentNumber_; - if (((from_bitField0_ & 0x00000400) == 0x00000400)) { - to_bitField0_ |= 0x00000080; - } - result.attemptNumber_ = attemptNumber_; + result.vertexId_ = vertexId_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto) { - return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto)other); + if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier) { + return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier)other); } else { super.mergeFrom(other); return this; } } - public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto other) { - if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto.getDefaultInstance()) return this; - if (other.hasFragmentIdentifierString()) { + public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier other) { + if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier.getDefaultInstance()) return this; + if (other.hasApplicationIdString()) { bitField0_ |= 0x00000001; - fragmentIdentifierString_ = other.fragmentIdentifierString_; + applicationIdString_ = other.applicationIdString_; onChanged(); } - if (other.hasDagName()) { - bitField0_ |= 0x00000002; - dagName_ = other.dagName_; - onChanged(); + if (other.hasAppAttemptNumber()) { + setAppAttemptNumber(other.getAppAttemptNumber()); } if (other.hasDagId()) { setDagId(other.getDagId()); } - if (other.hasVertexName()) { - bitField0_ |= 0x00000008; - vertexName_ = other.vertexName_; - onChanged(); - } - if (other.hasProcessorDescriptor()) { - mergeProcessorDescriptor(other.getProcessorDescriptor()); - } - if (inputSpecsBuilder_ == null) { - if (!other.inputSpecs_.isEmpty()) { - if (inputSpecs_.isEmpty()) { - inputSpecs_ = other.inputSpecs_; - bitField0_ = (bitField0_ & ~0x00000020); - } else { - ensureInputSpecsIsMutable(); - inputSpecs_.addAll(other.inputSpecs_); - } - onChanged(); - } - } else { - if (!other.inputSpecs_.isEmpty()) { - if (inputSpecsBuilder_.isEmpty()) { - inputSpecsBuilder_.dispose(); - inputSpecsBuilder_ = null; - inputSpecs_ = other.inputSpecs_; - bitField0_ = (bitField0_ & ~0x00000020); - inputSpecsBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? - getInputSpecsFieldBuilder() : null; - } else { - inputSpecsBuilder_.addAllMessages(other.inputSpecs_); - } - } - } - if (outputSpecsBuilder_ == null) { - if (!other.outputSpecs_.isEmpty()) { - if (outputSpecs_.isEmpty()) { - outputSpecs_ = other.outputSpecs_; - bitField0_ = (bitField0_ & ~0x00000040); - } else { - ensureOutputSpecsIsMutable(); - outputSpecs_.addAll(other.outputSpecs_); - } - onChanged(); - } - } else { - if (!other.outputSpecs_.isEmpty()) { - if (outputSpecsBuilder_.isEmpty()) { - outputSpecsBuilder_.dispose(); - outputSpecsBuilder_ = null; - outputSpecs_ = other.outputSpecs_; - bitField0_ = (bitField0_ & ~0x00000040); - outputSpecsBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? - getOutputSpecsFieldBuilder() : null; - } else { - outputSpecsBuilder_.addAllMessages(other.outputSpecs_); - } - } - } - if (groupedInputSpecsBuilder_ == null) { - if (!other.groupedInputSpecs_.isEmpty()) { - if (groupedInputSpecs_.isEmpty()) { - groupedInputSpecs_ = other.groupedInputSpecs_; - bitField0_ = (bitField0_ & ~0x00000080); - } else { - ensureGroupedInputSpecsIsMutable(); - groupedInputSpecs_.addAll(other.groupedInputSpecs_); - } - onChanged(); - } - } else { - if (!other.groupedInputSpecs_.isEmpty()) { - if (groupedInputSpecsBuilder_.isEmpty()) { - groupedInputSpecsBuilder_.dispose(); - groupedInputSpecsBuilder_ = null; - groupedInputSpecs_ = other.groupedInputSpecs_; - bitField0_ = (bitField0_ & ~0x00000080); - groupedInputSpecsBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? - getGroupedInputSpecsFieldBuilder() : null; - } else { - groupedInputSpecsBuilder_.addAllMessages(other.groupedInputSpecs_); - } - } - } - if (other.hasVertexParallelism()) { - setVertexParallelism(other.getVertexParallelism()); - } - if (other.hasFragmentNumber()) { - setFragmentNumber(other.getFragmentNumber()); - } - if (other.hasAttemptNumber()) { - setAttemptNumber(other.getAttemptNumber()); + if (other.hasVertexId()) { + setVertexId(other.getVertexId()); } this.mergeUnknownFields(other.getUnknownFields()); return this; @@ -4499,11 +3792,11 @@ public final class LlapDaemonProtocolProtos { com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto parsedMessage = null; + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentSpecProto) e.getUnfinishedMessage(); + parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { @@ -4514,170 +3807,129 @@ public final class LlapDaemonProtocolProtos { } private int bitField0_; - // optional string fragment_identifier_string = 1; - private java.lang.Object fragmentIdentifierString_ = ""; + // optional string application_id_string = 1; + private java.lang.Object applicationIdString_ = ""; /** - * optional string fragment_identifier_string = 1; + * optional string application_id_string = 1; */ - public boolean hasFragmentIdentifierString() { + public boolean hasApplicationIdString() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** - * optional string fragment_identifier_string = 1; + * optional string application_id_string = 1; */ - public java.lang.String getFragmentIdentifierString() { - java.lang.Object ref = fragmentIdentifierString_; + public java.lang.String getApplicationIdString() { + java.lang.Object ref = applicationIdString_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); - fragmentIdentifierString_ = s; + applicationIdString_ = s; return s; } else { return (java.lang.String) ref; } } /** - * optional string fragment_identifier_string = 1; + * optional string application_id_string = 1; */ public com.google.protobuf.ByteString - getFragmentIdentifierStringBytes() { - java.lang.Object ref = fragmentIdentifierString_; + getApplicationIdStringBytes() { + java.lang.Object ref = applicationIdString_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); - fragmentIdentifierString_ = b; + applicationIdString_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** - * optional string fragment_identifier_string = 1; + * optional string application_id_string = 1; */ - public Builder setFragmentIdentifierString( + public Builder setApplicationIdString( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; - fragmentIdentifierString_ = value; + applicationIdString_ = value; onChanged(); return this; } /** - * optional string fragment_identifier_string = 1; + * optional string application_id_string = 1; */ - public Builder clearFragmentIdentifierString() { + public Builder clearApplicationIdString() { bitField0_ = (bitField0_ & ~0x00000001); - fragmentIdentifierString_ = getDefaultInstance().getFragmentIdentifierString(); + applicationIdString_ = getDefaultInstance().getApplicationIdString(); onChanged(); return this; } /** - * optional string fragment_identifier_string = 1; + * optional string application_id_string = 1; */ - public Builder setFragmentIdentifierStringBytes( + public Builder setApplicationIdStringBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; - fragmentIdentifierString_ = value; + applicationIdString_ = value; onChanged(); return this; } - // optional string dag_name = 2; - private java.lang.Object dagName_ = ""; + // optional int32 app_attempt_number = 2; + private int appAttemptNumber_ ; /** - * optional string dag_name = 2; + * optional int32 app_attempt_number = 2; */ - public boolean hasDagName() { + public boolean hasAppAttemptNumber() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** - * optional string dag_name = 2; + * optional int32 app_attempt_number = 2; */ - public java.lang.String getDagName() { - java.lang.Object ref = dagName_; - if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - dagName_ = s; - return s; - } else { - return (java.lang.String) ref; - } + public int getAppAttemptNumber() { + return appAttemptNumber_; } /** - * optional string dag_name = 2; + * optional int32 app_attempt_number = 2; */ - public com.google.protobuf.ByteString - getDagNameBytes() { - java.lang.Object ref = dagName_; - if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - dagName_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } - } - /** - * optional string dag_name = 2; - */ - public Builder setDagName( - java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000002; - dagName_ = value; + public Builder setAppAttemptNumber(int value) { + bitField0_ |= 0x00000002; + appAttemptNumber_ = value; onChanged(); return this; } /** - * optional string dag_name = 2; + * optional int32 app_attempt_number = 2; */ - public Builder clearDagName() { + public Builder clearAppAttemptNumber() { bitField0_ = (bitField0_ & ~0x00000002); - dagName_ = getDefaultInstance().getDagName(); - onChanged(); - return this; - } - /** - * optional string dag_name = 2; - */ - public Builder setDagNameBytes( - com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000002; - dagName_ = value; + appAttemptNumber_ = 0; onChanged(); return this; } - // optional int32 dag_id = 11; + // optional int32 dag_id = 3; private int dagId_ ; /** - * optional int32 dag_id = 11; + * optional int32 dag_id = 3; */ public boolean hasDagId() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** - * optional int32 dag_id = 11; + * optional int32 dag_id = 3; */ public int getDagId() { return dagId_; } /** - * optional int32 dag_id = 11; + * optional int32 dag_id = 3; */ public Builder setDagId(int value) { bitField0_ |= 0x00000004; @@ -4686,7 +3938,7 @@ public final class LlapDaemonProtocolProtos { return this; } /** - * optional int32 dag_id = 11; + * optional int32 dag_id = 3; */ public Builder clearDagId() { bitField0_ = (bitField0_ & ~0x00000004); @@ -4695,1025 +3947,3552 @@ public final class LlapDaemonProtocolProtos { return this; } - // optional string vertex_name = 3; - private java.lang.Object vertexName_ = ""; + // optional int32 vertex_id = 4; + private int vertexId_ ; /** - * optional string vertex_name = 3; + * optional int32 vertex_id = 4; */ - public boolean hasVertexName() { + public boolean hasVertexId() { return ((bitField0_ & 0x00000008) == 0x00000008); } /** - * optional string vertex_name = 3; - */ - public java.lang.String getVertexName() { - java.lang.Object ref = vertexName_; - if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - vertexName_ = s; - return s; - } else { - return (java.lang.String) ref; - } - } - /** - * optional string vertex_name = 3; + * optional int32 vertex_id = 4; */ - public com.google.protobuf.ByteString - getVertexNameBytes() { - java.lang.Object ref = vertexName_; - if (ref instanceof String) { - com.google.protobuf.ByteString b = - com.google.protobuf.ByteString.copyFromUtf8( - (java.lang.String) ref); - vertexName_ = b; - return b; - } else { - return (com.google.protobuf.ByteString) ref; - } + public int getVertexId() { + return vertexId_; } /** - * optional string vertex_name = 3; + * optional int32 vertex_id = 4; */ - public Builder setVertexName( - java.lang.String value) { - if (value == null) { - throw new NullPointerException(); - } - bitField0_ |= 0x00000008; - vertexName_ = value; + public Builder setVertexId(int value) { + bitField0_ |= 0x00000008; + vertexId_ = value; onChanged(); return this; } /** - * optional string vertex_name = 3; + * optional int32 vertex_id = 4; */ - public Builder clearVertexName() { + public Builder clearVertexId() { bitField0_ = (bitField0_ & ~0x00000008); - vertexName_ = getDefaultInstance().getVertexName(); + vertexId_ = 0; onChanged(); return this; } - /** - * optional string vertex_name = 3; - */ - public Builder setVertexNameBytes( - com.google.protobuf.ByteString value) { - if (value == null) { - throw new NullPointerException(); + + // @@protoc_insertion_point(builder_scope:VertexIdentifier) + } + + static { + defaultInstance = new VertexIdentifier(true); + defaultInstance.initFields(); + } + + // @@protoc_insertion_point(class_scope:VertexIdentifier) } - bitField0_ |= 0x00000008; - vertexName_ = value; - onChanged(); - return this; - } - // optional .EntityDescriptorProto processor_descriptor = 4; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> processorDescriptorBuilder_; - /** - * optional .EntityDescriptorProto processor_descriptor = 4; - */ - public boolean hasProcessorDescriptor() { - return ((bitField0_ & 0x00000010) == 0x00000010); - } - /** - * optional .EntityDescriptorProto processor_descriptor = 4; - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getProcessorDescriptor() { - if (processorDescriptorBuilder_ == null) { - return processorDescriptor_; - } else { - return processorDescriptorBuilder_.getMessage(); - } - } - /** - * optional .EntityDescriptorProto processor_descriptor = 4; - */ - public Builder setProcessorDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) { - if (processorDescriptorBuilder_ == null) { - if (value == null) { - throw new NullPointerException(); - } - processorDescriptor_ = value; - onChanged(); - } else { - processorDescriptorBuilder_.setMessage(value); - } - bitField0_ |= 0x00000010; - return this; - } - /** - * optional .EntityDescriptorProto processor_descriptor = 4; - */ - public Builder setProcessorDescriptor( - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder builderForValue) { - if (processorDescriptorBuilder_ == null) { - processorDescriptor_ = builderForValue.build(); - onChanged(); - } else { - processorDescriptorBuilder_.setMessage(builderForValue.build()); - } - bitField0_ |= 0x00000010; - return this; - } - /** - * optional .EntityDescriptorProto processor_descriptor = 4; - */ - public Builder mergeProcessorDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) { - if (processorDescriptorBuilder_ == null) { - if (((bitField0_ & 0x00000010) == 0x00000010) && - processorDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) { - processorDescriptor_ = - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.newBuilder(processorDescriptor_).mergeFrom(value).buildPartial(); - } else { - processorDescriptor_ = value; - } - onChanged(); - } else { - processorDescriptorBuilder_.mergeFrom(value); - } - bitField0_ |= 0x00000010; - return this; - } - /** - * optional .EntityDescriptorProto processor_descriptor = 4; - */ - public Builder clearProcessorDescriptor() { - if (processorDescriptorBuilder_ == null) { - processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); - onChanged(); - } else { - processorDescriptorBuilder_.clear(); - } - bitField0_ = (bitField0_ & ~0x00000010); - return this; - } - /** - * optional .EntityDescriptorProto processor_descriptor = 4; - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder getProcessorDescriptorBuilder() { - bitField0_ |= 0x00000010; - onChanged(); - return getProcessorDescriptorFieldBuilder().getBuilder(); - } - /** - * optional .EntityDescriptorProto processor_descriptor = 4; - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getProcessorDescriptorOrBuilder() { - if (processorDescriptorBuilder_ != null) { - return processorDescriptorBuilder_.getMessageOrBuilder(); - } else { - return processorDescriptor_; - } - } - /** - * optional .EntityDescriptorProto processor_descriptor = 4; - */ - private com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> - getProcessorDescriptorFieldBuilder() { - if (processorDescriptorBuilder_ == null) { - processorDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilder< - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder>( - processorDescriptor_, - getParentForChildren(), - isClean()); - processorDescriptor_ = null; - } - return processorDescriptorBuilder_; - } + public interface SignableVertexSpecOrBuilder + extends com.google.protobuf.MessageOrBuilder { - // repeated .IOSpecProto input_specs = 5; - private java.util.List inputSpecs_ = - java.util.Collections.emptyList(); - private void ensureInputSpecsIsMutable() { - if (!((bitField0_ & 0x00000020) == 0x00000020)) { - inputSpecs_ = new java.util.ArrayList(inputSpecs_); - bitField0_ |= 0x00000020; - } - } + // optional string user = 1; + /** + * optional string user = 1; + */ + boolean hasUser(); + /** + * optional string user = 1; + */ + java.lang.String getUser(); + /** + * optional string user = 1; + */ + com.google.protobuf.ByteString + getUserBytes(); - private com.google.protobuf.RepeatedFieldBuilder< - org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> inputSpecsBuilder_; + // optional int64 signatureKeyId = 2; + /** + * optional int64 signatureKeyId = 2; + */ + boolean hasSignatureKeyId(); + /** + * optional int64 signatureKeyId = 2; + */ + long getSignatureKeyId(); - /** - * repeated .IOSpecProto input_specs = 5; - */ - public java.util.List getInputSpecsList() { - if (inputSpecsBuilder_ == null) { - return java.util.Collections.unmodifiableList(inputSpecs_); - } else { - return inputSpecsBuilder_.getMessageList(); - } - } - /** - * repeated .IOSpecProto input_specs = 5; - */ - public int getInputSpecsCount() { - if (inputSpecsBuilder_ == null) { - return inputSpecs_.size(); - } else { - return inputSpecsBuilder_.getCount(); - } - } - /** - * repeated .IOSpecProto input_specs = 5; - */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getInputSpecs(int index) { - if (inputSpecsBuilder_ == null) { - return inputSpecs_.get(index); - } else { - return inputSpecsBuilder_.getMessage(index); - } - } - /** - * repeated .IOSpecProto input_specs = 5; - */ - public Builder setInputSpecs( - int index, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto value) { - if (inputSpecsBuilder_ == null) { + // optional .VertexIdentifier vertexIdentifier = 3; + /** + * optional .VertexIdentifier vertexIdentifier = 3; + */ + boolean hasVertexIdentifier(); + /** + * optional .VertexIdentifier vertexIdentifier = 3; + */ + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifier getVertexIdentifier(); + /** + * optional .VertexIdentifier vertexIdentifier = 3; + */ + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexIdentifierOrBuilder getVertexIdentifierOrBuilder(); + + // optional string dag_name = 4; + /** + * optional string dag_name = 4; + * + *
+     * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
+     * 
+ */ + boolean hasDagName(); + /** + * optional string dag_name = 4; + * + *
+     * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
+     * 
+ */ + java.lang.String getDagName(); + /** + * optional string dag_name = 4; + * + *
+     * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
+     * 
+ */ + com.google.protobuf.ByteString + getDagNameBytes(); + + // optional string vertex_name = 5; + /** + * optional string vertex_name = 5; + */ + boolean hasVertexName(); + /** + * optional string vertex_name = 5; + */ + java.lang.String getVertexName(); + /** + * optional string vertex_name = 5; + */ + com.google.protobuf.ByteString + getVertexNameBytes(); + + // optional string token_identifier = 6; + /** + * optional string token_identifier = 6; + * + *
+     * The core vertex stuff 
+     * 
+ */ + boolean hasTokenIdentifier(); + /** + * optional string token_identifier = 6; + * + *
+     * The core vertex stuff 
+     * 
+ */ + java.lang.String getTokenIdentifier(); + /** + * optional string token_identifier = 6; + * + *
+     * The core vertex stuff 
+     * 
+ */ + com.google.protobuf.ByteString + getTokenIdentifierBytes(); + + // optional .EntityDescriptorProto processor_descriptor = 7; + /** + * optional .EntityDescriptorProto processor_descriptor = 7; + */ + boolean hasProcessorDescriptor(); + /** + * optional .EntityDescriptorProto processor_descriptor = 7; + */ + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getProcessorDescriptor(); + /** + * optional .EntityDescriptorProto processor_descriptor = 7; + */ + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getProcessorDescriptorOrBuilder(); + + // repeated .IOSpecProto input_specs = 8; + /** + * repeated .IOSpecProto input_specs = 8; + */ + java.util.List + getInputSpecsList(); + /** + * repeated .IOSpecProto input_specs = 8; + */ + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getInputSpecs(int index); + /** + * repeated .IOSpecProto input_specs = 8; + */ + int getInputSpecsCount(); + /** + * repeated .IOSpecProto input_specs = 8; + */ + java.util.List + getInputSpecsOrBuilderList(); + /** + * repeated .IOSpecProto input_specs = 8; + */ + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getInputSpecsOrBuilder( + int index); + + // repeated .IOSpecProto output_specs = 9; + /** + * repeated .IOSpecProto output_specs = 9; + */ + java.util.List + getOutputSpecsList(); + /** + * repeated .IOSpecProto output_specs = 9; + */ + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getOutputSpecs(int index); + /** + * repeated .IOSpecProto output_specs = 9; + */ + int getOutputSpecsCount(); + /** + * repeated .IOSpecProto output_specs = 9; + */ + java.util.List + getOutputSpecsOrBuilderList(); + /** + * repeated .IOSpecProto output_specs = 9; + */ + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getOutputSpecsOrBuilder( + int index); + + // repeated .GroupInputSpecProto grouped_input_specs = 10; + /** + * repeated .GroupInputSpecProto grouped_input_specs = 10; + */ + java.util.List + getGroupedInputSpecsList(); + /** + * repeated .GroupInputSpecProto grouped_input_specs = 10; + */ + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getGroupedInputSpecs(int index); + /** + * repeated .GroupInputSpecProto grouped_input_specs = 10; + */ + int getGroupedInputSpecsCount(); + /** + * repeated .GroupInputSpecProto grouped_input_specs = 10; + */ + java.util.List + getGroupedInputSpecsOrBuilderList(); + /** + * repeated .GroupInputSpecProto grouped_input_specs = 10; + */ + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder getGroupedInputSpecsOrBuilder( + int index); + + // optional int32 vertex_parallelism = 11; + /** + * optional int32 vertex_parallelism = 11; + * + *
+     * An internal field required for Tez.
+     * 
+ */ + boolean hasVertexParallelism(); + /** + * optional int32 vertex_parallelism = 11; + * + *
+     * An internal field required for Tez.
+     * 
+ */ + int getVertexParallelism(); + } + /** + * Protobuf type {@code SignableVertexSpec} + * + *
+   * The part of SubmitWork that can be signed 
+   * 
+ */ + public static final class SignableVertexSpec extends + com.google.protobuf.GeneratedMessage + implements SignableVertexSpecOrBuilder { + // Use SignableVertexSpec.newBuilder() to construct. + private SignableVertexSpec(com.google.protobuf.GeneratedMessage.Builder builder) { + super(builder); + this.unknownFields = builder.getUnknownFields(); + } + private SignableVertexSpec(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } + + private static final SignableVertexSpec defaultInstance; + public static SignableVertexSpec getDefaultInstance() { + return defaultInstance; + } + + public SignableVertexSpec getDefaultInstanceForType() { + return defaultInstance; + } + + private final com.google.protobuf.UnknownFieldSet unknownFields; + @java.lang.Override + public final com.google.protobuf.UnknownFieldSet + getUnknownFields() { + return this.unknownFields; + } + private SignableVertexSpec( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + initFields(); + int mutable_bitField0_ = 0; + c