Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id 32EFA200C03 for ; Sat, 21 Jan 2017 15:56:37 +0100 (CET) Received: by cust-asf.ponee.io (Postfix) id 26F39160B4A; Sat, 21 Jan 2017 14:56:37 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 256AC160B3C for ; Sat, 21 Jan 2017 15:56:35 +0100 (CET) Received: (qmail 51022 invoked by uid 500); 21 Jan 2017 14:56:35 -0000 Mailing-List: contact issues-help@hive.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@hive.apache.org Delivered-To: mailing list issues@hive.apache.org Received: (qmail 51013 invoked by uid 99); 21 Jan 2017 14:56:35 -0000 Received: from pnap-us-west-generic-nat.apache.org (HELO spamd2-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 21 Jan 2017 14:56:35 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd2-us-west.apache.org (ASF Mail Server at spamd2-us-west.apache.org) with ESMTP id E1D151A04D7 for ; Sat, 21 Jan 2017 14:56:34 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd2-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: 0.002 X-Spam-Level: X-Spam-Status: No, score=0.002 tagged_above=-999 required=6.31 tests=[KAM_BADIPHTTP=2, KAM_LAZY_DOMAIN_SECURITY=1, NORMAL_HTTP_TO_IP=0.001, RP_MATCHES_RCVD=-2.999] autolearn=disabled Received: from mx1-lw-eu.apache.org ([10.40.0.8]) by localhost (spamd2-us-west.apache.org [10.40.0.9]) (amavisd-new, port 10024) with ESMTP id DWd0TNmfnZez for ; Sat, 21 Jan 2017 14:56:32 +0000 (UTC) Received: from mailrelay1-us-west.apache.org (mailrelay1-us-west.apache.org [209.188.14.139]) by mx1-lw-eu.apache.org (ASF Mail Server at mx1-lw-eu.apache.org) with ESMTP id 6E1225F286 for ; Sat, 21 Jan 2017 14:56:31 +0000 (UTC) Received: from jira-lw-us.apache.org (unknown [207.244.88.139]) by mailrelay1-us-west.apache.org (ASF Mail Server at mailrelay1-us-west.apache.org) with ESMTP id 58346E0236 for ; Sat, 21 Jan 2017 14:56:28 +0000 (UTC) Received: from jira-lw-us.apache.org (localhost [127.0.0.1]) by jira-lw-us.apache.org (ASF Mail Server at jira-lw-us.apache.org) with ESMTP id 2F2A525287 for ; Sat, 21 Jan 2017 14:56:27 +0000 (UTC) Date: Sat, 21 Jan 2017 14:56:27 +0000 (UTC) From: "Hive QA (JIRA)" To: issues@hive.apache.org Message-ID: In-Reply-To: References: Subject: [jira] [Commented] (HIVE-15439) Support INSERT OVERWRITE for internal druid datasources. MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit X-JIRA-FingerPrint: 30527f35849b9dde25b450d4833f0394 archived-at: Sat, 21 Jan 2017 14:56:37 -0000 [ https://issues.apache.org/jira/browse/HIVE-15439?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15833015#comment-15833015 ] Hive QA commented on HIVE-15439: -------------------------------- Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12848660/HIVE-15439.5.patch {color:red}ERROR:{color} -1 due to build exiting with an error Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/3103/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/3103/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-3103/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Tests exited with: NonZeroExitCodeException Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit status 1 and output '+ date '+%Y-%m-%d %T.%3N' 2017-01-21 14:53:01.912 + [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]] + export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64 + JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64 + export PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games + PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games + export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m ' + ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m ' + export 'MAVEN_OPTS=-Xmx1g ' + MAVEN_OPTS='-Xmx1g ' + cd /data/hiveptest/working/ + tee /data/hiveptest/logs/PreCommit-HIVE-Build-3103/source-prep.txt + [[ false == \t\r\u\e ]] + mkdir -p maven ivy + [[ git = \s\v\n ]] + [[ git = \g\i\t ]] + [[ -z master ]] + [[ -d apache-github-source-source ]] + [[ ! -d apache-github-source-source/.git ]] + [[ ! -d apache-github-source-source ]] + date '+%Y-%m-%d %T.%3N' 2017-01-21 14:53:01.915 + cd apache-github-source-source + git fetch origin + git reset --hard HEAD HEAD is now at d9343f6 HIVE-15544 : Support scalar subqueries (Vineet Garg via Ashutosh Chauhan) + git clean -f -d + git checkout master Already on 'master' Your branch is up-to-date with 'origin/master'. + git reset --hard origin/master HEAD is now at d9343f6 HIVE-15544 : Support scalar subqueries (Vineet Garg via Ashutosh Chauhan) + git merge --ff-only origin/master Already up-to-date. + date '+%Y-%m-%d %T.%3N' 2017-01-21 14:53:02.796 + patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh + patchFilePath=/data/hiveptest/working/scratch/build.patch + [[ -f /data/hiveptest/working/scratch/build.patch ]] + chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh + /data/hiveptest/working/scratch/smart-apply-patch.sh /data/hiveptest/working/scratch/build.patch Going to apply patch with: patch -p0 patching file accumulo-handler/src/test/results/positive/accumulo_queries.q.out patching file accumulo-handler/src/test/results/positive/accumulo_single_sourced_multi_insert.q.out patching file druid-handler/pom.xml patching file druid-handler/src/java/org/apache/hadoop/hive/druid/DruidStorageHandler.java patching file druid-handler/src/test/org/apache/hadoop/hive/druid/DruidStorageHandlerTest.java patching file druid-handler/src/test/org/apache/hadoop/hive/druid/TestDerbyConnector.java patching file druid-handler/src/test/org/apache/hadoop/hive/ql/io/DruidRecordWriterTest.java patching file hbase-handler/src/test/results/positive/hbase_queries.q.out patching file hbase-handler/src/test/results/positive/hbase_single_sourced_multi_insert.q.out patching file hbase-handler/src/test/results/positive/hbasestats.q.out patching file metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaHookV2.java patching file metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java patching file metastore/src/java/org/apache/hadoop/hive/metastore/IMetaStoreClient.java patching file ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java patching file ql/src/java/org/apache/hadoop/hive/ql/optimizer/SortedDynPartitionTimeGranularityOptimizer.java patching file ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java Hunk #7 succeeded at 10770 (offset 3 lines). Hunk #8 succeeded at 10795 (offset 3 lines). Hunk #9 succeeded at 10917 (offset 3 lines). patching file ql/src/java/org/apache/hadoop/hive/ql/plan/DDLWork.java patching file ql/src/java/org/apache/hadoop/hive/ql/plan/InsertTableDesc.java + [[ maven == \m\a\v\e\n ]] + rm -rf /data/hiveptest/working/maven/org/apache/hive + mvn -B clean install -DskipTests -T 4 -q -Dmaven.repo.local=/data/hiveptest/working/maven ANTLR Parser Generator Version 3.5.2 Output file /data/hiveptest/working/apache-github-source-source/metastore/target/generated-sources/antlr3/org/apache/hadoop/hive/metastore/parser/FilterParser.java does not exist: must build /data/hiveptest/working/apache-github-source-source/metastore/src/java/org/apache/hadoop/hive/metastore/parser/Filter.g org/apache/hadoop/hive/metastore/parser/Filter.g DataNucleus Enhancer (version 4.1.6) for API "JDO" DataNucleus Enhancer : Classpath >> /usr/share/maven/boot/plexus-classworlds-2.x.jar ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MDatabase ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MFieldSchema ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MType ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MTable ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MConstraint ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MSerDeInfo ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MOrder ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MColumnDescriptor ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MStringList ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MStorageDescriptor ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MPartition ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MIndex ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MRole ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MRoleMap ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MGlobalPrivilege ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MDBPrivilege ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MTablePrivilege ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MPartitionPrivilege ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MTableColumnPrivilege ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MPartitionColumnPrivilege ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MPartitionEvent ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MMasterKey ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MDelegationToken ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MTableColumnStatistics ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MPartitionColumnStatistics ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MVersionTable ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MResourceUri ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MFunction ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MNotificationLog ENHANCED (Persistable) : org.apache.hadoop.hive.metastore.model.MNotificationNextId DataNucleus Enhancer completed with success for 30 classes. Timings : input=206 ms, enhance=272 ms, total=478 ms. Consult the log for full details ANTLR Parser Generator Version 3.5.2 Output file /data/hiveptest/working/apache-github-source-source/ql/target/generated-sources/antlr3/org/apache/hadoop/hive/ql/parse/HiveLexer.java does not exist: must build /data/hiveptest/working/apache-github-source-source/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveLexer.g org/apache/hadoop/hive/ql/parse/HiveLexer.g Output file /data/hiveptest/working/apache-github-source-source/ql/target/generated-sources/antlr3/org/apache/hadoop/hive/ql/parse/HiveParser.java does not exist: must build /data/hiveptest/working/apache-github-source-source/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g org/apache/hadoop/hive/ql/parse/HiveParser.g Generating vector expression code Generating vector expression test code [ERROR] COMPILATION ERROR : [ERROR] /data/hiveptest/working/apache-github-source-source/druid-handler/src/test/org/apache/hadoop/hive/druid/DruidStorageHandlerTest.java:[159,47] no suitable constructor found for DruidStorageHandler(org.apache.hadoop.hive.druid.TestDerbyConnector,io.druid.indexer.SQLMetadataStorageUpdaterJobHandler,io.druid.metadata.MetadataStorageTablesConfig) constructor org.apache.hadoop.hive.druid.DruidStorageHandler.DruidStorageHandler() is not applicable (actual and formal argument lists differ in length) constructor org.apache.hadoop.hive.druid.DruidStorageHandler.DruidStorageHandler(io.druid.metadata.SQLMetadataConnector,io.druid.indexer.SQLMetadataStorageUpdaterJobHandler,io.druid.metadata.MetadataStorageTablesConfig,com.metamx.http.client.HttpClient) is not applicable (actual and formal argument lists differ in length) SLF4J: Class path contains multiple SLF4J bindings. SLF4J: Found binding in [jar:file:/data/hiveptest/working/maven/org/apache/logging/log4j/log4j-slf4j-impl/2.6.2/log4j-slf4j-impl-2.6.2.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found binding in [jar:file:/data/hiveptest/working/maven/org/slf4j/slf4j-log4j12/1.7.10/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory] ERROR StatusLogger No log4j2 configuration file found. Using default configuration: logging only errors to the console. [ERROR] Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:3.1:testCompile (default-testCompile) on project hive-druid-handler: Compilation failure [ERROR] /data/hiveptest/working/apache-github-source-source/druid-handler/src/test/org/apache/hadoop/hive/druid/DruidStorageHandlerTest.java:[159,47] no suitable constructor found for DruidStorageHandler(org.apache.hadoop.hive.druid.TestDerbyConnector,io.druid.indexer.SQLMetadataStorageUpdaterJobHandler,io.druid.metadata.MetadataStorageTablesConfig) [ERROR] constructor org.apache.hadoop.hive.druid.DruidStorageHandler.DruidStorageHandler() is not applicable [ERROR] (actual and formal argument lists differ in length) [ERROR] constructor org.apache.hadoop.hive.druid.DruidStorageHandler.DruidStorageHandler(io.druid.metadata.SQLMetadataConnector,io.druid.indexer.SQLMetadataStorageUpdaterJobHandler,io.druid.metadata.MetadataStorageTablesConfig,com.metamx.http.client.HttpClient) is not applicable [ERROR] (actual and formal argument lists differ in length) [ERROR] -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn -rf :hive-druid-handler + exit 1 ' {noformat} This message is automatically generated. ATTACHMENT ID: 12848660 - PreCommit-HIVE-Build > Support INSERT OVERWRITE for internal druid datasources. > -------------------------------------------------------- > > Key: HIVE-15439 > URL: https://issues.apache.org/jira/browse/HIVE-15439 > Project: Hive > Issue Type: Sub-task > Components: Druid integration > Affects Versions: 2.2.0 > Reporter: slim bouguerra > Assignee: slim bouguerra > Attachments: HIVE-15439.3.patch, HIVE-15439.4.patch, HIVE-15439.5.patch, HIVE-15439.patch, HIVE-15439.patch, HIVE-15439.patch, HIVE-15439.patch, HIVE-15439.patch, HIVE-15439.patch > > > Add support for SQL statement INSERT OVERWRITE TABLE druid_internal_table. > In order to add this support will need to add new post insert hook to update the druid metadata. Creation of the segment will be the same as CTAS. > -- This message was sent by Atlassian JIRA (v6.3.4#6332)