Return-Path: X-Original-To: apmail-hive-user-archive@www.apache.org Delivered-To: apmail-hive-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 213BE1062A for ; Fri, 27 Sep 2013 23:21:21 +0000 (UTC) Received: (qmail 72232 invoked by uid 500); 27 Sep 2013 23:21:19 -0000 Delivered-To: apmail-hive-user-archive@hive.apache.org Received: (qmail 72171 invoked by uid 500); 27 Sep 2013 23:21:19 -0000 Mailing-List: contact user-help@hive.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hive.apache.org Delivered-To: mailing list user@hive.apache.org Received: (qmail 72163 invoked by uid 99); 27 Sep 2013 23:21:19 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 27 Sep 2013 23:21:19 +0000 X-ASF-Spam-Status: No, hits=1.7 required=5.0 tests=FREEMAIL_ENVFROM_END_DIGIT,HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of xguo27@gmail.com designates 209.85.223.173 as permitted sender) Received: from [209.85.223.173] (HELO mail-ie0-f173.google.com) (209.85.223.173) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 27 Sep 2013 23:21:12 +0000 Received: by mail-ie0-f173.google.com with SMTP id ar20so5392145iec.4 for ; Fri, 27 Sep 2013 16:20:52 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:from:date:message-id:subject:to:content-type; bh=U9gBL8MfceMEToyDh7rR/q/daifBrnpwWg/MytTxCWk=; b=xJxGrpHBOXmDG4VKXhsFYNXv+rokBmPzpfAr6NB4xKki0JWLuZFj9sklOJkO7TVA5e 7yO8CqEoSjL+B09/qYaAgsrBbaSFQcCJd5i27ZCZIiRKS9dIDbGimpa9IylmyWb2E4sr 23QP1Nh/cQr0iRAMON2sodSGCYUw+wJ4R0v9sL4viMWb3SHOwACLZ74qf9X0e6gAn0H1 tnhQqfWcoXImvy/5YExZNLet/gBXwrYeuFxS6HOuTPkRTOnI5X6hTVWVkO1ixkFq2EeP 0eU1WWYk06tdLBZiUqfbDyhqV32p46n7ABMZA6dZlxHJ7HnzMvP+2GB4ooZSi7AUYwor Sjnw== X-Received: by 10.50.108.71 with SMTP id hi7mr4415050igb.59.1380324052034; Fri, 27 Sep 2013 16:20:52 -0700 (PDT) MIME-Version: 1.0 Received: by 10.64.241.103 with HTTP; Fri, 27 Sep 2013 16:20:31 -0700 (PDT) From: Xiu Guo Date: Fri, 27 Sep 2013 16:20:31 -0700 Message-ID: Subject: Error while running junit test for hive in eclipse To: user@hive.apache.org Content-Type: multipart/alternative; boundary=089e0149c4a0cbeb5d04e765bc0c X-Virus-Checked: Checked by ClamAV on apache.org --089e0149c4a0cbeb5d04e765bc0c Content-Type: text/plain; charset=ISO-8859-1 I compiled Hive trunk source code using: ant clean package eclipse-files and run a junit test on command line just fine: ant test -Dtestcase=TestCliDriver -Dqfile=auto_join25.q -Dmodule=ql -Dtest.silent=false However, when imported into eclipse and run the same test, it throws following error: 13/09/26 13:44:41 INFO mr.ExecDriver: Executing: /home/xiu/git/hiveTrunk/testutils/hadoop jar /home/xiu/git/hiveTrunk/build/ql/hive-exec-0.13.0-SNAPSHOT.jar org.apache.hadoop.hive.ql.exec.mr.ExecDriver -plan file:/home/xiu/git/hiveTrunk/build/ql/localscratchdir/hive_2013-09-26_13-44-40_635_1327282357933361987-1/-local-10004/plan.xml -jobconffile file:/home/xiu/git/hiveTrunk/build/ql/localscratchdir/hive_2013-09-26_13-44-40_635_1327282357933361987-1/-local-10003/jobconf.xml which: no /home/xiu/git/hiveTrunk/testutils/hadoop in ((null)) dirname: missing operand Try `dirname --help' for more information. Exception in thread "main" java.lang.NoClassDefFoundError: org.apache.hadoop.hive.metastore.api.AlreadyExistsException at java.lang.J9VMInternals.verifyImpl(Native Method) at java.lang.J9VMInternals.verify(J9VMInternals.java:72) at java.lang.J9VMInternals.initialize(J9VMInternals.java:134) at org.apache.hadoop.hive.ql.plan.PartitionDesc.getDeserializer(PartitionDesc.java:112) at org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getSplits(CombineHiveInputFormat.java:310) at org.apache.hadoop.mapred.JobClient.writeOldSplits(JobClient.java:1051) at org.apache.hadoop.mapred.JobClient.writeSplits(JobClient.java:1043) at org.apache.hadoop.mapred.JobClient.access$700(JobClient.java:179) at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:959) at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:912) at java.security.AccessController.doPrivileged(AccessController.java:362) at javax.security.auth.Subject.doAs(Subject.java:573) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1149) at org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:912) at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:886) at org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:425) at org.apache.hadoop.hive.ql.exec.mr.ExecDriver.main(ExecDriver.java:725) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:60) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:37) at java.lang.reflect.Method.invoke(Method.java:611) at org.apache.hadoop.util.RunJar.main(RunJar.java:156) Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.hive.metastore.api.AlreadyExistsException at java.net.URLClassLoader.findClass(URLClassLoader.java:434) at java.lang.ClassLoader.loadClass(ClassLoader.java:672) at java.lang.ClassLoader.loadClass(ClassLoader.java:638) ... 22 more 13/09/26 13:44:43 ERROR exec.Task: Execution failed with exit status: 1 13/09/26 13:44:43 ERROR exec.Task: Obtaining error information 13/09/26 13:44:43 ERROR exec.Task: Task failed! Task ID: Stage-1 Logs: 13/09/26 13:44:43 ERROR mr.ExecDriver: Execution failed with exit status: 1 It had been fine until I updated to the latest Trunk code two days ago. I looked up, the AlreadyExistsException.java is there before build, AlreadyExistsException.class is there after build and \*metastore*.jar is there too. Could anyone please tell me where else should I look into? Thank you! --089e0149c4a0cbeb5d04e765bc0c Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable
I compiled Hive trunk source c= ode using:

=A0=A0=A0 ant clean package eclipse-files

and run = a junit test on command line just fine:

=A0=A0=A0 ant test -Dtestcas= e=3DTestCliDriver -Dqfile=3Dauto_join25.q -Dmodule=3Dql -Dtest.silent=3Dfal= se

However, when imported into eclipse and run the same test, it throws fo= llowing error:



=A0=A0=A0 13/09/26 13:44:41 INFO mr.ExecDrive= r: Executing: /home/xiu/git/hiveTrunk/testutils/hadoop jar /home/xiu/git/hi= veTrunk/build/ql/hive-exec-0.13.0-SNAPSHOT.jar org.apache.hadoop.hive.ql.ex= ec.mr.ExecDriver=A0 -plan file:/home/xiu/git/hiveTrunk/build/ql/localscratc= hdir/hive_2013-09-26_13-44-40_635_1327282357933361987-1/-local-10004/plan.x= ml=A0=A0 -jobconffile file:/home/xiu/git/hiveTrunk/build/ql/localscratchdir= /hive_2013-09-26_13-44-40_635_1327282357933361987-1/-local-10003/jobconf.xm= l
=A0=A0=A0 which: no /home/xiu/git/hiveTrunk/testutils/hadoop in ((null))=A0=A0=A0 dirname: missing operand
=A0=A0=A0 Try `dirname --help' f= or more information.
=A0=A0=A0 Exception in thread "main" java= .lang.NoClassDefFoundError: org.apache.hadoop.hive.metastore.api.AlreadyExi= stsException
=A0=A0=A0 =A0=A0=A0 at java.lang.J9VMInternals.verifyImpl(Native Method)=A0=A0=A0 =A0=A0=A0 at java.lang.J9VMInternals.verify(J9VMInternals.java:7= 2)
=A0=A0=A0 =A0=A0=A0 at java.lang.J9VMInternals.initialize(J9VMInterna= ls.java:134)
=A0=A0=A0 =A0=A0=A0 at org.apache.hadoop.hive.ql.plan.Parti= tionDesc.getDeserializer(PartitionDesc.java:112)
=A0=A0=A0 =A0=A0=A0 at org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.= getSplits(CombineHiveInputFormat.java:310)
=A0=A0=A0 =A0=A0=A0 at org.ap= ache.hadoop.mapred.JobClient.writeOldSplits(JobClient.java:1051)
=A0=A0= =A0 =A0=A0=A0 at org.apache.hadoop.mapred.JobClient.writeSplits(JobClient.j= ava:1043)
=A0=A0=A0 =A0=A0=A0 at org.apache.hadoop.mapred.JobClient.access$700(JobCli= ent.java:179)
=A0=A0=A0 =A0=A0=A0 at org.apache.hadoop.mapred.JobClient$= 2.run(JobClient.java:959)
=A0=A0=A0 =A0=A0=A0 at org.apache.hadoop.mapre= d.JobClient$2.run(JobClient.java:912)
=A0=A0=A0 =A0=A0=A0 at java.security.AccessController.doPrivileged(AccessCo= ntroller.java:362)
=A0=A0=A0 =A0=A0=A0 at javax.security.auth.Subject.do= As(Subject.java:573)
=A0=A0=A0 =A0=A0=A0 at org.apache.hadoop.security.U= serGroupInformation.doAs(UserGroupInformation.java:1149)
=A0=A0=A0 =A0=A0=A0 at org.apache.hadoop.mapred.JobClient.submitJobInternal= (JobClient.java:912)
=A0=A0=A0 =A0=A0=A0 at org.apache.hadoop.mapred.Job= Client.submitJob(JobClient.java:886)
=A0=A0=A0 =A0=A0=A0 at org.apache.h= adoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:425)
=A0=A0=A0 =A0=A0=A0 at org.apache.hadoop.hive.ql.exec.mr.ExecDriver.main(Ex= ecDriver.java:725)
=A0=A0=A0 =A0=A0=A0 at sun.reflect.NativeMethodAccess= orImpl.invoke0(Native Method)
=A0=A0=A0 =A0=A0=A0 at sun.reflect.NativeM= ethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:60)
=A0=A0=A0 =A0=A0=A0 at sun.reflect.DelegatingMethodAccessorImpl.invoke(Dele= gatingMethodAccessorImpl.java:37)
=A0=A0=A0 =A0=A0=A0 at java.lang.refle= ct.Method.invoke(Method.java:611)
=A0=A0=A0 =A0=A0=A0 at org.apache.hado= op.util.RunJar.main(RunJar.java:156)
=A0=A0=A0 Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.hi= ve.metastore.api.AlreadyExistsException
=A0=A0=A0 at java.net.URLClassLo= ader.findClass(URLClassLoader.java:434)
=A0=A0=A0 at java.lang.ClassLoad= er.loadClass(ClassLoader.java:672)
=A0=A0=A0 at java.lang.ClassLoader.loadClass(ClassLoader.java:638)
=A0= =A0=A0 ... 22 more
=A0=A0=A0 13/09/26 13:44:43 ERROR exec.Task: Executio= n failed with exit status: 1
=A0=A0=A0 13/09/26 13:44:43 ERROR exec.Task= : Obtaining error information
=A0=A0=A0 13/09/26 13:44:43 ERROR exec.Task:
=A0=A0=A0 Task failed!
= =A0=A0=A0 Task ID:
=A0=A0=A0=A0=A0 Stage-1

=A0=A0=A0 Logs:
=A0=A0=A0 13/09/26 13:44:43 ERROR mr.ExecDriver: Execution failed with exi= t status: 1

It had been fine until I updated to the latest Trunk= code two days ago.

I loo= ked up, the AlreadyExistsException.java is there before build, AlreadyExist= sException.class is there after build and \*metastore*.jar is there too.
Could anyone please tell me where else should I look into?

Thank= you!
--089e0149c4a0cbeb5d04e765bc0c--