hive-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From vipul sharma <sharmavipulw...@gmail.com>
Subject Re: Errors in hive queries
Date Wed, 27 Apr 2011 23:54:38 GMT
This got resolved. Some of the clients were not upgraded to CHD3 and were
running old version.

On Wed, Apr 27, 2011 at 2:27 PM, vipul sharma <sharmavipulwork@gmail.com>wrote:

> We have a hadoop/hive cluster which is using cloudera's distribution.
> Metastore is stored in mysql and all the relevant drivers are in classpath
> and in conf files. While running queries on hive I am getting these errors
>
> $ hive -hiveconf hive.root.logger=INFO,console
> Hive history file=/tmp/vipul/hive_job_log_vipul_201104271409_1310845475.txt
> 11/04/27 14:09:01 INFO exec.HiveHistory: Hive history
> file=/tmp/vipul/hive_job_log_vipul_201104271409_1310845475.txt
> hive> select time from requests_stat_min;
> 11/04/27 14:15:38 INFO parse.ParseDriver: Parsing command: select time from
> requests_stat_min
> 11/04/27 14:15:38 INFO parse.ParseDriver: Parse Completed
> 11/04/27 14:15:39 INFO parse.SemanticAnalyzer: Starting Semantic Analysis
> 11/04/27 14:15:39 INFO parse.SemanticAnalyzer: Completed phase 1 of
> Semantic Analysis
> 11/04/27 14:15:39 INFO parse.SemanticAnalyzer: Get metadata for source
> tables
> 11/04/27 14:15:39 INFO metastore.HiveMetaStore: 0: Opening raw store with
> implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
> 11/04/27 14:15:39 INFO metastore.ObjectStore: ObjectStore, initialize
> called
> 11/04/27 14:15:39 ERROR DataNucleus.Plugin: Bundle "org.eclipse.jdt.core"
> requires "org.eclipse.core.resources" but it cannot be resolved.
> 11/04/27 14:15:39 ERROR DataNucleus.Plugin: Bundle "org.eclipse.jdt.core"
> requires "org.eclipse.core.runtime" but it cannot be resolved.
> 11/04/27 14:15:39 ERROR DataNucleus.Plugin: Bundle "org.eclipse.jdt.core"
> requires "org.eclipse.text" but it cannot be resolved.
> 11/04/27 14:15:40 INFO metastore.ObjectStore: Initialized ObjectStore
> 11/04/27 14:15:40 INFO metastore.HiveMetaStore: 0: get_table : db=default
> tbl=requests_stat_min
> 11/04/27 14:15:41 INFO hive.log: DDL: struct requests_stat_min { string
> time, i64 request_count, i64 get_count, i64 post_count, i64 unique_orders,
> i64 unique_events, i64 start_orders, i32 max_response_time, i32
> min_response_time, i32 avg_response_time, i32 max_response_size, i32
> min_response_size, i32 avg_response_size, i64 unique_referrers}
> 11/04/27 14:15:41 INFO parse.SemanticAnalyzer: Get metadata for subqueries
> 11/04/27 14:15:41 INFO parse.SemanticAnalyzer: Get metadata for destination
> tables
> 11/04/27 14:15:41 INFO parse.SemanticAnalyzer: Completed getting MetaData
> in Semantic Analysis
> 11/04/27 14:15:41 INFO ppd.OpProcFactory: Processing for FS(2)
> 11/04/27 14:15:41 INFO ppd.OpProcFactory: Processing for SEL(1)
> 11/04/27 14:15:41 INFO ppd.OpProcFactory: Processing for TS(0)
> 11/04/27 14:15:41 INFO metastore.HiveMetaStore: 0: get_partition_names :
> db=default tbl=requests_stat_min
> 11/04/27 14:15:41 INFO metastore.HiveMetaStore: 0: get_partition :
> db=default tbl=requests_stat_min
> 11/04/27 14:15:41 INFO metastore.HiveMetaStore: 0: get_partition :
> db=default tbl=requests_stat_min
> 11/04/27 14:15:42 INFO metastore.HiveMetaStore: 0: get_partition :
> db=default tbl=requests_stat_min
> 11/04/27 14:15:42 INFO metastore.HiveMetaStore: 0: get_partition :
> db=default tbl=requests_stat_min
> 11/04/27 14:15:42 INFO metastore.HiveMetaStore: 0: get_partition :
> db=default tbl=requests_stat_min
> 11/04/27 14:15:42 INFO metastore.HiveMetaStore: 0: get_partition :
> db=default tbl=requests_stat_min
> 11/04/27 14:15:42 INFO metastore.HiveMetaStore: 0: get_partition :
> db=default tbl=requests_stat_min
> 11/04/27 14:15:42 INFO metastore.HiveMetaStore: 0: get_partition :
> db=default tbl=requests_stat_min
> 11/04/27 14:15:42 INFO metastore.HiveMetaStore: 0: get_partition :
> db=default tbl=requests_stat_min
> 11/04/27 14:15:42 INFO metastore.HiveMetaStore: 0: get_partition :
> db=default tbl=requests_stat_min
> 11/04/27 14:15:42 INFO hive.log: DDL: struct requests_stat_min { string
> time, i64 request_count, i64 get_count, i64 post_count, i64 unique_orders,
> i64 unique_events, i64 start_orders, i32 max_response_time, i32
> min_response_time, i32 avg_response_time, i32 max_response_size, i32
> min_response_size, i32 avg_response_size, i64 unique_referrers}
> 11/04/27 14:15:42 INFO hive.log: DDL: struct requests_stat_min { string
> time, i64 request_count, i64 get_count, i64 post_count, i64 unique_orders,
> i64 unique_events, i64 start_orders, i32 max_response_time, i32
> min_response_time, i32
> 11/04/27 14:15:42 INFO parse.SemanticAnalyzer: Completed plan generation
> 11/04/27 14:15:42 INFO ql.Driver: Semantic Analysis Completed
> 11/04/27 14:15:42 INFO ql.Driver: Starting command: select time from
> requests_stat_min
> Total MapReduce jobs = 1
> 11/04/27 14:15:42 INFO ql.Driver: Total MapReduce jobs = 1
> Launching Job 1 out of 1
> 11/04/27 14:15:42 INFO ql.Driver: Launching Job 1 out of 1
> Number of reduce tasks is set to 0 since there's no reduce operator
> 11/04/27 14:15:42 INFO exec.ExecDriver: Number of reduce tasks is set to 0
> since there's no reduce operator
> FAILED: Unknown exception : null
> 11/04/27 14:15:42 ERROR ql.Driver: FAILED: Unknown exception : null
> java.lang.NullPointerException
>     at java.util.Hashtable.put(Hashtable.java:394)
>     at java.util.Properties.setProperty(Properties.java:143)
>     at org.apache.hadoop.conf.Configuration.set(Configuration.java:460)
>     at org.apache.hadoop.hive.conf.HiveConf.setVar(HiveConf.java:293)
>     at
> org.apache.hadoop.hive.ql.exec.ExecDriver.execute(ExecDriver.java:505)
>     at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:100)
>     at
> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:64)
>     at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:572)
>     at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:452)
>     at org.apache.hadoop.hive.ql.Driver.runCommand(Driver.java:314)
>     at org.apache.hadoop.hive.ql.Driver.run(Driver.java:302)
>     at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:123)
>     at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:181)
>     at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:287)
>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>     at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>     at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>     at java.lang.reflect.Method.invoke(Method.java:597)
>     at org.apache.hadoop.util.RunJar.main(RunJar.java:186)
>
>
> I am assuming this has to do with metastore but I cant figure out whats
> wrong. The way we run query is via a remote client. Any help is greatly
> appreciated!
>
> --
> Vipul Sharma
> sharmavipul AT gmail DOT com
>



-- 
Vipul Sharma
sharmavipul AT gmail DOT com

Mime
View raw message