ambari-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Hudson (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (AMBARI-12220) HiveServer2 query fail after RU
Date Thu, 02 Jul 2015 05:56:04 GMT

    [ https://issues.apache.org/jira/browse/AMBARI-12220?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14611521#comment-14611521
] 

Hudson commented on AMBARI-12220:
---------------------------------

SUCCESS: Integrated in Ambari-branch-2.1 #164 (See [https://builds.apache.org/job/Ambari-branch-2.1/164/])
AMBARI-12220: HiveServer2 query fail after RU (jluniya) (jluniya: http://git-wip-us.apache.org/repos/asf?p=ambari.git&a=commit&h=0807bf493d6e47f3fb059574af5a2bce751bf33c)
* ambari-server/src/test/java/org/apache/ambari/server/upgrade/UpgradeCatalog210Test.java
* ambari-server/src/main/java/org/apache/ambari/server/upgrade/UpgradeCatalog210.java


> HiveServer2 query fail after RU
> -------------------------------
>
>                 Key: AMBARI-12220
>                 URL: https://issues.apache.org/jira/browse/AMBARI-12220
>             Project: Ambari
>          Issue Type: Bug
>          Components: ambari-server
>    Affects Versions: 2.1.0
>            Reporter: Jayush Luniya
>            Assignee: Jayush Luniya
>            Priority: Blocker
>             Fix For: 2.1.0
>
>         Attachments: AMBARI-12220.patch
>
>
> After Rolling upgrade, HiveServer2 query fail with the message:
> {code}
> ERROR : Job Submission failed with exception 'java.io.FileNotFoundException(File file:/usr/hdp/current/hive-webhcat/share/hcatalog/hive-hcatalog-core-0.14.0.2.2.6.0-2800.jar
does not exist)'
> java.io.FileNotFoundException: File file:/usr/hdp/current/hive-webhcat/share/hcatalog/hive-hcatalog-core-0.14.0.2.2.6.0-2800.jar
does not exist
> 	at org.apache.hadoop.fs.RawLocalFileSystem.deprecatedGetFileStatus(RawLocalFileSystem.java:606)
> 	at org.apache.hadoop.fs.RawLocalFileSystem.getFileLinkStatusInternal(RawLocalFileSystem.java:819)
> 	at org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:596)
> 	at org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:421)
> 	at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:337)
> 	at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:289)
> 	at org.apache.hadoop.mapreduce.JobResourceUploader.copyRemoteFiles(JobResourceUploader.java:203)
> 	at org.apache.hadoop.mapreduce.JobResourceUploader.uploadFiles(JobResourceUploader.java:128)
> 	at org.apache.hadoop.mapreduce.JobSubmitter.copyAndConfigureFiles(JobSubmitter.java:95)
> 	at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:190)
> 	at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290)
> 	at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287)
> 	at java.security.AccessController.doPrivileged(Native Method)
> 	at javax.security.auth.Subject.doAs(Subject.java:415)
> 	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
> 	at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287)
> 	at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:575)
> 	at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:570)
> 	at java.security.AccessController.doPrivileged(Native Method)
> 	at javax.security.auth.Subject.doAs(Subject.java:415)
> 	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
> 	at org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:570)
> 	at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:561)
> 	at org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:431)
> 	at org.apache.hadoop.hive.ql.exec.mr.MapRedTask.execute(MapRedTask.java:137)
> 	at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160)
> 	at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:88)
> 	at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1653)
> 	at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1412)
> 	at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1195)
> 	at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1059)
> 	at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1054)
> 	at org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:154)
> 	at org.apache.hive.service.cli.operation.SQLOperation.access$100(SQLOperation.java:71)
> 	at org.apache.hive.service.cli.operation.SQLOperation$1$1.run(SQLOperation.java:206)
> 	at java.security.AccessController.doPrivileged(Native Method)
> 	at javax.security.auth.Subject.doAs(Subject.java:415)
> 	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
> 	at org.apache.hive.service.cli.operation.SQLOperation$1.run(SQLOperation.java:218)
> 	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
> 	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> 	at java.lang.Thread.run(Thread.java:745)
> {code}
> Look into the HiveServer2 process command line:
> /usr/jdk64/jdk1.7.0_67/bin/java -Xmx1024m -Dhdp.version=2.3.0.0-2434 -Djava.net.preferIPv4Stack=true
-Dhdp.version=2.3.0.0-2434 -Dhadoop.log.dir=/var/log/hadoop/hive -Dhadoop.log.file=hadoop.log
-Dhadoop.home.dir=/usr/hdp/2.3.0.0-2434/hadoop -Dhadoop.id.str=hive -Dhadoop.root.logger=INFO,console
-Djava.library.path=:/usr/hdp/2.3.0.0-2434/hadoop/lib/native/Linux-amd64-64:/usr/hdp/2.3.0.0-2434/hadoop/lib/native
-Dhadoop.policy.file=hadoop-policy.xml -Djava.net.preferIPv4Stack=true -Xmx1024m -XX:MaxPermSize=512m
-Xmx1024m -Dhadoop.security.logger=INFO,NullAppender org.apache.hadoop.util.RunJar /usr/hdp/2.3.0.0-2434/hive/lib/hive-service-1.2.1.2.3.0.0-2434.jar
org.apache.hive.service.server.HiveServer2 --hiveconf hive.aux.jars.path=file:///usr/hdp/current/hive-webhcat/share/hcatalog/hive-hcatalog-core-0.14.0.2.2.6.0-2800.jar,file:///usr/hdp/current/hive-webhcat/share/hcatalog/hive-hcatalog-core.jar,file:///usr/hdp/current/hive-webhcat/share/hcatalog/hive-hcatalog-pig-adapter-0.14.0.2.2.6.0-2800.jar,file:///usr/hdp/current/hive-webhcat/share/hcatalog/hive-hcatalog-pig-adapter.jar,file:///usr/hdp/current/hive-webhcat/share/hcatalog/hive-hcatalog-server-extensions-0.14.0.2.2.6.0-2800.jar,file:///usr/hdp/current/hive-webhcat/share/hcatalog/hive-hcatalog-server-extensions.jar,file:///usr/hdp/current/hive-webhcat/share/hcatalog/hive-hcatalog-streaming-0.14.0.2.2.6.0-2800.jar,file:///usr/hdp/current/hive-webhcat/share/hcatalog/hive-hcatalog-streaming.jar
-hiveconf hive.metastore.uris=  -hiveconf hive.log.file=hiveserver2.log -hiveconf hive.log.dir=/var/log/hive
> The version in hcatalog jars is the old one. The jars do not exist after upgrade.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Mime
View raw message