hadoop-hive-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Edward Capriolo <edlinuxg...@gmail.com>
Subject Re: null pointer in 3.0
Date Sat, 18 Apr 2009 14:40:42 GMT
I was using default settings. Are  'APP' and 'mine' the default?

jpox...
javax.jdo.PersistenceManagerFactoryClass=org.jpox.PersistenceManagerFactoryImpl
org.jpox.autoCreateSchema=false
org.jpox.validateTables=false
org.jpox.validateColumns=false
org.jpox.validateConstraints=false
org.jpox.storeManagerType=rdbms
org.jpox.autoCreateSchema=true
org.jpox.autoStartMechanismMode=checked
org.jpox.transactionIsolation=read_committed
javax.jdo.option.DetachAllOnCommit=true
javax.jdo.option.NontransactionalRead=true
javax.jdo.option.ConnectionDriverName=org.apache.derby.jdbc.ClientDriver
javax.jdo.option.ConnectionURL=jdbc:derby://hadoop1.jointhegrid.local:1527/metastore_db;create=true
javax.jdo.option.ConnectionUserName=APP
javax.jdo.option.ConnectionPassword=mine
org.jpox.cache.level2=true
org.jpox.cache.level2.type=SOFT

hive-site

<configuration>
<property>
  <name>hive.metastore.local</name>
  <value>true</value>
  <description>controls whether to connect to remove metastore server
or open a new metastore server in Hive Client JVM</description>
</property>

<property>
  <name>javax.jdo.option.ConnectionURL</name>
  <value>jdbc:derby://hadoop1.jointhegrid.local:1527/metastore_db;create=true</value>
  <description>JDBC connect string for a JDBC metastore</description>
</property>

<property>
  <name>javax.jdo.option.ConnectionDriverName</name>
  <value>org.apache.derby.jdbc.ClientDriver</value>
  <description>Driver class name for a JDBC metastore</description>
</property>
</configuration>


On Fri, Apr 17, 2009 at 11:24 PM, Prasad Chakka <pchakka@facebook.com> wrote:
> Did you change the password parameter (for the metastore db) from the default? If so
is it empty? Because, the branch and trunk are working for me and that is the only case I
think could cause problems and want to make sure that is the cased before doing a checkin.
>
> Prasad
>
>
> ________________________________
> From: Edward Capriolo <edlinuxguru@gmail.com>
> Reply-To: <hive-dev@hadoop.apache.org>
> Date: Fri, 17 Apr 2009 19:40:49 -0700
> To: <hive-dev@hadoop.apache.org>
> Subject: Re: null pointer in 3.0
>
> I setup using the local meta store as well as a remote metastore using
> the derby client. Both of them were failing. After applying your patch
> my trunk version is now working. I only tested one query.
>
> I would suggest making a release candidate, because the 3.0 release is
> not functional.
>
>
> On Fri, Apr 17, 2009 at 8:56 PM, Ashish Thusoo <athusoo@facebook.com> wrote:
>> Shoudl we build a new release candidate for this? Is there a reasonable workaround?
>>
>> Ashish
>>
>> -----Original Message-----
>> From: Prasad Chakka [mailto:pchakka@facebook.com]
>> Sent: Friday, April 17, 2009 3:30 PM
>> To: hive-dev@hadoop.apache.org
>> Subject: Re: null pointer in 3.0
>>
>> Hi Edward,
>>
>> Can you try to apply this patch https://issues.apache.org/jira/secure/attachment/12405820/tmp1.patch
and let me know it works?
>>
>> What type of metastore setup are you using here?
>>
>> Thanks,
>> Prasad
>>
>>
>> ________________________________
>> From: Edward Capriolo <edlinuxguru@gmail.com>
>> Reply-To: <hive-dev@hadoop.apache.org>
>> Date: Fri, 17 Apr 2009 14:40:13 -0700
>> To: <hive-dev@hadoop.apache.org>
>> Subject: null pointer in 3.0
>>
>> Hey all,
>> I checked out 3.0 and build it against hadoop 0.18.3. Technically I am running against
the Cloudera version witch is patched 0.18.3. However I have also checked out the trunk and
built against 0.19.0 and seem to be having the same issue.
>>
>> I have a two column table straight selects work but any time I try a select (distinct),
or anything that involves map reduce the job fails.
>>
>> 09/04/17 17:05:45 ERROR exec.ExecDriver: Ended Job =
>> job_200904101611_0018 with exception
>> 'java.lang.NullPointerException(null)'
>> java.lang.NullPointerException
>>        at java.util.Hashtable.put(Hashtable.java:394)
>>        at java.util.Properties.setProperty(Properties.java:143)
>>        at org.apache.hadoop.conf.Configuration.set(Configuration.java:300)
>>        at org.apache.hadoop.hive.ql.exec.ExecDriver.execute(ExecDriver.java:398)
>>        at org.apache.hadoop.hive.ql.Driver.run(Driver.java:245)
>>        at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:174)
>>        at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:207)
>>        at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:306)
>>        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>        at java.lang.reflect.Method.invoke(Method.java:597)
>>        at org.apache.hadoop.util.RunJar.main(RunJar.java:155)
>>        at org.apache.hadoop.mapred.JobShell.run(JobShell.java:54)
>>        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
>>        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79)
>>        at org.apache.hadoop.mapred.JobShell.main(JobShell.java:68)
>>
>> FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.ExecDriver
>>
>> Any ideas?
>>
>>
>
>

Mime
View raw message