hive-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Babak Bastan <babak...@gmail.com>
Subject Re: Error while Creating Table in Hive
Date Wed, 06 Jun 2012 20:12:26 GMT
no one error:
i.e if I run this one

*hostname --fqdn*

 with the condition that I send to you :

*127.0.0.1       localhost*
*#127.0.0.1      ubuntu.ubuntu-domain    ubuntu*
*# The following lines are desirable for IPv6 capable hosts*
*#::1     ip6-localhost ip6-loopback*
*#fe00::0 ip6-localnet*
*#ff00::0 ip6-mcastprefix*
*#ff02::1 ip6-allnodes*
*#ff02::2 ip6-allrouters*

I get this error:

*hostname: Name or service not known*

Or in the second step by this command:

*babak@ubuntu:~/Downloads/hadoop/bin$ start-hdfs.sh*

these lines of error:


mkdir: kann Verzeichnis „/home/babak/Downloads/hadoop/bin/../logs“ nicht
anlegen: Keine Berechtigung
starting namenode, logging to
/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-namenode-ubuntu.out
/home/babak/Downloads/hadoop/bin/hadoop-daemon.sh: Zeile 117:
/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-namenode-ubuntu.out:
Datei oder Verzeichnis nicht gefunden
head:
„/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-namenode-ubuntu.out“
kann nicht zum Lesen geöffnet werden: Datei oder Verzeichnis nicht gefunden
localhost: mkdir: kann Verzeichnis
„/home/babak/Downloads/hadoop/bin/../logs“ nicht anlegen: Keine Berechtigung
localhost: starting datanode, logging to
/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-datanode-ubuntu.out
localhost: /home/babak/Downloads/hadoop/bin/hadoop-daemon.sh: Zeile 117:
/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-datanode-ubuntu.out:
Datei oder Verzeichnis nicht gefunden
localhost: head:
„/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-datanode-ubuntu.out“
kann nicht zum Lesen geöffnet werden: Datei oder Verzeichnis nicht gefunden
localhost: mkdir: kann Verzeichnis
„/home/babak/Downloads/hadoop/bin/../logs“ nicht anlegen: Keine Berechtigung
localhost: starting secondarynamenode, logging to
/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-secondarynamenode-ubuntu.out
localhost: /home/babak/Downloads/hadoop/bin/hadoop-daemon.sh: Zeile 117:
/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-secondarynamenode-ubuntu.out:
Datei oder Verzeichnis nicht gefunden
localhost: head:
„/home/babak/Downloads/hadoop/bin/../logs/hadoop-babak-secondarynamenode-ubuntu.out“
kann nicht zum Lesen geöffnet werden: Datei oder Verzeichnis nicht gefunden

they said no permision to make logs in this
path:/home/babak/Downloads/hadoop/bin/../logs

 and generally I cant create a table in hive and get this one:

FAILED: Error in metadata: MetaException(message:Got exception:
java.io.FileNotFoundException File file:/user/hive/warehouse/test does
not exist.)
FAILED: Execution Error, return code 1 from
org.apache.hadoop.hive.ql.exec.DDLTask

On Wed, Jun 6, 2012 at 10:02 PM, shashwat shriparv <
dwivedishashwat@gmail.com> wrote:

> whats the error babak ???
>
>
> On Thu, Jun 7, 2012 at 1:25 AM, Babak Bastan <babakbsn@gmail.com> wrote:
>
>> What the hell is that?I see no log folder there
>>
>>
>> On Wed, Jun 6, 2012 at 9:41 PM, Mohammad Tariq <dontariq@gmail.com>wrote:
>>
>>> go to your HADOOP_HOME i.e your hadoop directory(that includes bin,
>>> conf etc)..you can find logs directory there..
>>>
>>> Regards,
>>>     Mohammad Tariq
>>>
>>>
>>> On Thu, Jun 7, 2012 at 1:09 AM, Babak Bastan <babakbsn@gmail.com> wrote:
>>> > hoe can I get my log mohammad?
>>> >
>>> >
>>> > On Wed, Jun 6, 2012 at 9:36 PM, Mohammad Tariq <dontariq@gmail.com>
>>> wrote:
>>> >>
>>> >> could you post your logs???that would help me in understanding the
>>> >> problem properly.
>>> >>
>>> >> Regards,
>>> >>     Mohammad Tariq
>>> >>
>>> >>
>>> >> On Thu, Jun 7, 2012 at 1:02 AM, Babak Bastan <babakbsn@gmail.com>
>>> wrote:
>>> >> > Thank you very much mohamad for your attention.I followed the steps
>>> but
>>> >> > the
>>> >> > error is the same as the last time.
>>> >> > and there is my hosts file:
>>> >> >
>>> >> > 127.0.0.1       localhost
>>> >> > #127.0.0.1      ubuntu.ubuntu-domain    ubuntu
>>> >> >
>>> >> >
>>> >> > # The following lines are desirable for IPv6 capable hosts
>>> >> >
>>> >> > #::1     ip6-localhost ip6-loopback
>>> >> > #fe00::0 ip6-localnet
>>> >> > #ff00::0 ip6-mcastprefix
>>> >> > #ff02::1 ip6-allnodes
>>> >> > #ff02::2 ip6-allrouters
>>> >> >
>>> >> > but no effect :(
>>> >> >
>>> >> > On Wed, Jun 6, 2012 at 8:25 PM, Mohammad Tariq <dontariq@gmail.com>
>>> >> > wrote:
>>> >> >>
>>> >> >> also change the permissions of these directories to 777.
>>> >> >>
>>> >> >> Regards,
>>> >> >>     Mohammad Tariq
>>> >> >>
>>> >> >>
>>> >> >> On Wed, Jun 6, 2012 at 11:54 PM, Mohammad Tariq <
>>> dontariq@gmail.com>
>>> >> >> wrote:
>>> >> >> > create a directory "/home/username/hdfs" (or at some place
of
>>> your
>>> >> >> > choice)..inside this hdfs directory create three sub directories
>>> -
>>> >> >> > name, data, and temp, then follow these steps :
>>> >> >> >
>>> >> >> > add following properties in your core-site.xml -
>>> >> >> >
>>> >> >> > <property>
>>> >> >> >          <name>fs.default.name</name>
>>> >> >> >          <value>hdfs://localhost:9000/</value>
>>> >> >> >        </property>
>>> >> >> >
>>> >> >> >        <property>
>>> >> >> >          <name>hadoop.tmp.dir</name>
>>> >> >> >          <value>/home/mohammad/hdfs/temp</value>
>>> >> >> >        </property>
>>> >> >> >
>>> >> >> > then add following two properties in your hdfs-site.xml
-
>>> >> >> >
>>> >> >> > <property>
>>> >> >> >                <name>dfs.replication</name>
>>> >> >> >                <value>1</value>
>>> >> >> >        </property>
>>> >> >> >
>>> >> >> >        <property>
>>> >> >> >                <name>dfs.name.dir</name>
>>> >> >> >                <value>/home/mohammad/hdfs/name</value>
>>> >> >> >        </property>
>>> >> >> >
>>> >> >> >        <property>
>>> >> >> >                <name>dfs.data.dir</name>
>>> >> >> >                <value>/home/mohammad/hdfs/data</value>
>>> >> >> >        </property>
>>> >> >> >
>>> >> >> > finally add this property in your mapred-site.xml -
>>> >> >> >
>>> >> >> >       <property>
>>> >> >> >          <name>mapred.job.tracker</name>
>>> >> >> >          <value>hdfs://localhost:9001</value>
>>> >> >> >        </property>
>>> >> >> >
>>> >> >> > NOTE: you can give any name to these directories of your
choice,
>>> just
>>> >> >> > keep in mind you have to give same names as values of
>>> >> >> >           above specified properties in your configuration
files.
>>> >> >> > (give full path of these directories, not just the name
of the
>>> >> >> > directory)
>>> >> >> >
>>> >> >> > After this  follow the steps provided in the previous
reply.
>>> >> >> >
>>> >> >> > Regards,
>>> >> >> >     Mohammad Tariq
>>> >> >> >
>>> >> >> >
>>> >> >> > On Wed, Jun 6, 2012 at 11:42 PM, Babak Bastan <
>>> babakbsn@gmail.com>
>>> >> >> > wrote:
>>> >> >> >> thank's Mohammad
>>> >> >> >>
>>> >> >> >> with this command:
>>> >> >> >>
>>> >> >> >> babak@ubuntu:~/Downloads/hadoop/bin$ hadoop namenode
-format
>>> >> >> >>
>>> >> >> >> this is my output:
>>> >> >> >>
>>> >> >> >> 12/06/06 20:05:20 INFO namenode.NameNode: STARTUP_MSG:
>>> >> >> >> /************************************************************
>>> >> >> >> STARTUP_MSG: Starting NameNode
>>> >> >> >> STARTUP_MSG:   host = ubuntu/127.0.1.1
>>> >> >> >> STARTUP_MSG:   args = [-format]
>>> >> >> >> STARTUP_MSG:   version = 0.20.2
>>> >> >> >> STARTUP_MSG:   build =
>>> >> >> >>
>>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20
>>> >> >> >> -r
>>> >> >> >> 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34
UTC 2010
>>> >> >> >> ************************************************************/
>>> >> >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem:
>>> >> >> >>
>>> >> >> >>
>>> fsOwner=babak,babak,adm,dialout,cdrom,plugdev,lpadmin,admin,sambashare
>>> >> >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem:
>>> supergroup=supergroup
>>> >> >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem:
>>> >> >> >> isPermissionEnabled=true
>>> >> >> >> 12/06/06 20:05:20 INFO common.Storage: Image file
of size 95
>>> saved
>>> >> >> >> in 0
>>> >> >> >> seconds.
>>> >> >> >> 12/06/06 20:05:20 INFO common.Storage: Storage directory
>>> >> >> >> /tmp/hadoop-babak/dfs/name has been successfully formatted.
>>> >> >> >> 12/06/06 20:05:20 INFO namenode.NameNode: SHUTDOWN_MSG:
>>> >> >> >> /************************************************************
>>> >> >> >> SHUTDOWN_MSG: Shutting down NameNode at ubuntu/127.0.1.1
>>> >> >> >> ************************************************************/
>>> >> >> >>
>>> >> >> >> by this command:
>>> >> >> >>
>>> >> >> >> babak@ubuntu:~/Downloads/hadoop/bin$ start-dfs.sh
>>> >> >> >>
>>> >> >> >> this is the out put
>>> >> >> >>
>>> >> >> >> mkdir: kann Verzeichnis
>>> „/home/babak/Downloads/hadoop/bin/../logs“
>>> >> >> >> nicht
>>> >> >> >> anlegen: Keine Berechtigung
>>> >> >> >>
>>> >> >> >> this out put(it's in german and it means no right
to make this
>>> >> >> >> folder)
>>> >> >> >>
>>> >> >> >>
>>> >> >> >> On Wed, Jun 6, 2012 at 7:59 PM, Mohammad Tariq <
>>> dontariq@gmail.com>
>>> >> >> >> wrote:
>>> >> >> >>>
>>> >> >> >>> once we are done with the configuration, we need
to format the
>>> file
>>> >> >> >>> system..use this command to do that-
>>> >> >> >>> bin/hadoop namenode -format
>>> >> >> >>>
>>> >> >> >>> after this, hadoop daemon processes should be
started using
>>> >> >> >>> following
>>> >> >> >>> commands -
>>> >> >> >>> bin/start-dfs.sh (it'll start NN & DN)
>>> >> >> >>> bin/start-mapred.sh (it'll start JT & TT)
>>> >> >> >>>
>>> >> >> >>> after this use jps to check if everything is alright
or point
>>> your
>>> >> >> >>> browser to localhost:50070..if you further find
any problem
>>> provide
>>> >> >> >>> us
>>> >> >> >>> with the error logs..:)
>>> >> >> >>>
>>> >> >> >>> Regards,
>>> >> >> >>>     Mohammad Tariq
>>> >> >> >>>
>>> >> >> >>>
>>> >> >> >>> On Wed, Jun 6, 2012 at 11:22 PM, Babak Bastan
<
>>> babakbsn@gmail.com>
>>> >> >> >>> wrote:
>>> >> >> >>> > were you able to format hdfs properly???
>>> >> >> >>> > I did'nt get your question,Do you mean HADOOP_HOME?
or where
>>> did
>>> >> >> >>> > I
>>> >> >> >>> > install
>>> >> >> >>> > Hadoop?
>>> >> >> >>> >
>>> >> >> >>> > On Wed, Jun 6, 2012 at 7:49 PM, Mohammad
Tariq
>>> >> >> >>> > <dontariq@gmail.com>
>>> >> >> >>> > wrote:
>>> >> >> >>> >>
>>> >> >> >>> >> if you are getting only this, it means
your hadoop is not
>>> >> >> >>> >> running..were you able to format hdfs
properly???
>>> >> >> >>> >>
>>> >> >> >>> >> Regards,
>>> >> >> >>> >>     Mohammad Tariq
>>> >> >> >>> >>
>>> >> >> >>> >>
>>> >> >> >>> >> On Wed, Jun 6, 2012 at 11:17 PM, Babak
Bastan
>>> >> >> >>> >> <babakbsn@gmail.com>
>>> >> >> >>> >> wrote:
>>> >> >> >>> >> > Hi MohammadmI irun jps in my shel
I can see this result:
>>> >> >> >>> >> > 2213 Jps
>>> >> >> >>> >> >
>>> >> >> >>> >> >
>>> >> >> >>> >> > On Wed, Jun 6, 2012 at 7:44 PM,
Mohammad Tariq
>>> >> >> >>> >> > <dontariq@gmail.com>
>>> >> >> >>> >> > wrote:
>>> >> >> >>> >> >>
>>> >> >> >>> >> >> you can also use "jps" command
at your shell to see
>>> whether
>>> >> >> >>> >> >> Hadoop
>>> >> >> >>> >> >> processes are running or not.
>>> >> >> >>> >> >>
>>> >> >> >>> >> >> Regards,
>>> >> >> >>> >> >>     Mohammad Tariq
>>> >> >> >>> >> >>
>>> >> >> >>> >> >>
>>> >> >> >>> >> >> On Wed, Jun 6, 2012 at 11:12
PM, Mohammad Tariq
>>> >> >> >>> >> >> <dontariq@gmail.com>
>>> >> >> >>> >> >> wrote:
>>> >> >> >>> >> >> > Hi Babak,
>>> >> >> >>> >> >> >
>>> >> >> >>> >> >> >  You have to type it in
you web browser..Hadoop
>>> provides us
>>> >> >> >>> >> >> > a
>>> >> >> >>> >> >> > web
>>> >> >> >>> >> >> > GUI
>>> >> >> >>> >> >> > that not only allows us
to browse through the file
>>> system,
>>> >> >> >>> >> >> > but
>>> >> >> >>> >> >> > to
>>> >> >> >>> >> >> > download the files as well..Apart
from that it also
>>> >> >> >>> >> >> > provides a
>>> >> >> >>> >> >> > web
>>> >> >> >>> >> >> > GUI
>>> >> >> >>> >> >> > that can be used to see
the status of Jobtracker and
>>> >> >> >>> >> >> > Tasktracker..When
>>> >> >> >>> >> >> > you run a Hive or Pig job
or a Mapreduce job, you can
>>> point
>>> >> >> >>> >> >> > your
>>> >> >> >>> >> >> > browser to http://localhost:50030
to see the status
>>> and
>>> >> >> >>> >> >> > logs
>>> >> >> >>> >> >> > of
>>> >> >> >>> >> >> > your
>>> >> >> >>> >> >> > job.
>>> >> >> >>> >> >> >
>>> >> >> >>> >> >> > Regards,
>>> >> >> >>> >> >> >     Mohammad Tariq
>>> >> >> >>> >> >> >
>>> >> >> >>> >> >> >
>>> >> >> >>> >> >> > On Wed, Jun 6, 2012 at
8:28 PM, Babak Bastan
>>> >> >> >>> >> >> > <babakbsn@gmail.com>
>>> >> >> >>> >> >> > wrote:
>>> >> >> >>> >> >> >> Thank you shashwat
for the answer,
>>> >> >> >>> >> >> >> where should I type
http://localhost:50070?
>>> >> >> >>> >> >> >> I typed here: hive>http://localhost:50070
but
>>> nothing as
>>> >> >> >>> >> >> >> result
>>> >> >> >>> >> >> >>
>>> >> >> >>> >> >> >>
>>> >> >> >>> >> >> >> On Wed, Jun 6, 2012
at 3:32 PM, shashwat shriparv
>>> >> >> >>> >> >> >> <dwivedishashwat@gmail.com>
wrote:
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>> first type http://localhost:50070
whether this is
>>> opening
>>> >> >> >>> >> >> >>> or
>>> >> >> >>> >> >> >>> not
>>> >> >> >>> >> >> >>> and
>>> >> >> >>> >> >> >>> check
>>> >> >> >>> >> >> >>> how many nodes
are available, check some of the
>>> hadoop
>>> >> >> >>> >> >> >>> shell
>>> >> >> >>> >> >> >>> commands
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>> from
>>> http://hadoop.apache.org/common/docs/r0.18.3/hdfs_shell.html
>>> >> >> >>> >> >> >>> run
>>> >> >> >>> >> >> >>> example mapreduce
task on hadoop take example from
>>> here
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>> :
>>> http://www.michael-noll.com/blog/2011/04/09/benchmarking-and-stress-testing-an-hadoop-cluster-with-terasort-testdfsio-nnbench-mrbench/
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>> if all the above
you can do sucessfully means hadoop
>>> is
>>> >> >> >>> >> >> >>> configured
>>> >> >> >>> >> >> >>> correctly
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>> Regards
>>> >> >> >>> >> >> >>> Shashwat
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>> On Wed, Jun 6,
2012 at 1:30 AM, Babak Bastan
>>> >> >> >>> >> >> >>> <babakbsn@gmail.com>
>>> >> >> >>> >> >> >>> wrote:
>>> >> >> >>> >> >> >>>>
>>> >> >> >>> >> >> >>>> no I'm not
working on CDH.Is there a way to test if
>>> my
>>> >> >> >>> >> >> >>>> Hadoop
>>> >> >> >>> >> >> >>>> works
>>> >> >> >>> >> >> >>>> fine
>>> >> >> >>> >> >> >>>> or not?
>>> >> >> >>> >> >> >>>>
>>> >> >> >>> >> >> >>>>
>>> >> >> >>> >> >> >>>> On Tue, Jun
5, 2012 at 9:55 PM, Bejoy KS
>>> >> >> >>> >> >> >>>> <bejoy_ks@yahoo.com>
>>> >> >> >>> >> >> >>>> wrote:
>>> >> >> >>> >> >> >>>>>
>>> >> >> >>> >> >> >>>>> Hi Babak
>>> >> >> >>> >> >> >>>>>
>>> >> >> >>> >> >> >>>>> You gotta
follow those instructions in the apace
>>> site
>>> >> >> >>> >> >> >>>>> to
>>> >> >> >>> >> >> >>>>> set
>>> >> >> >>> >> >> >>>>> up
>>> >> >> >>> >> >> >>>>> hadoop
>>> >> >> >>> >> >> >>>>> from scratch
and ensure that hdfs is working
>>> first. You
>>> >> >> >>> >> >> >>>>> should
>>> >> >> >>> >> >> >>>>> be
>>> >> >> >>> >> >> >>>>> able to
>>> >> >> >>> >> >> >>>>> read and
write files to hdfs before you do your
>>> next
>>> >> >> >>> >> >> >>>>> steps.
>>> >> >> >>> >> >> >>>>>
>>> >> >> >>> >> >> >>>>> Are you
on CDH or apache distribution of hadoop?
>>> If it
>>> >> >> >>> >> >> >>>>> is
>>> >> >> >>> >> >> >>>>> CDH
>>> >> >> >>> >> >> >>>>> there
>>> >> >> >>> >> >> >>>>> are
>>> >> >> >>> >> >> >>>>> detailed
instructions on Cloudera web site.
>>> >> >> >>> >> >> >>>>>
>>> >> >> >>> >> >> >>>>> Regards
>>> >> >> >>> >> >> >>>>> Bejoy KS
>>> >> >> >>> >> >> >>>>>
>>> >> >> >>> >> >> >>>>> Sent from
handheld, please excuse typos.
>>> >> >> >>> >> >> >>>>> ________________________________
>>> >> >> >>> >> >> >>>>> From: Babak
Bastan <babakbsn@gmail.com>
>>> >> >> >>> >> >> >>>>> Date: Tue,
5 Jun 2012 21:30:22 +0200
>>> >> >> >>> >> >> >>>>> To: <user@hive.apache.org>
>>> >> >> >>> >> >> >>>>> ReplyTo:
user@hive.apache.org
>>> >> >> >>> >> >> >>>>> Subject:
Re: Error while Creating Table in Hive
>>> >> >> >>> >> >> >>>>>
>>> >> >> >>> >> >> >>>>> @Bejoy:
I set the fs.default.name in the
>>> core-site.xml
>>> >> >> >>> >> >> >>>>> and
>>> >> >> >>> >> >> >>>>> I
>>> >> >> >>> >> >> >>>>> did
>>> >> >> >>> >> >> >>>>> all
>>> >> >> >>> >> >> >>>>> of
>>> >> >> >>> >> >> >>>>> thing that
was mentioned in the reference but no
>>> effect
>>> >> >> >>> >> >> >>>>>
>>> >> >> >>> >> >> >>>>> On Tue,
Jun 5, 2012 at 8:43 PM, Babak Bastan
>>> >> >> >>> >> >> >>>>> <babakbsn@gmail.com>
>>> >> >> >>> >> >> >>>>> wrote:
>>> >> >> >>> >> >> >>>>>>
>>> >> >> >>> >> >> >>>>>> Ok
sorry but that was my Mistake .I thought it
>>> works
>>> >> >> >>> >> >> >>>>>> but
>>> >> >> >>> >> >> >>>>>> no.
>>> >> >> >>> >> >> >>>>>> I wrote
the command without ; and then I think It
>>> >> >> >>> >> >> >>>>>> works
>>> >> >> >>> >> >> >>>>>> but
>>> >> >> >>> >> >> >>>>>> with
>>> >> >> >>> >> >> >>>>>> ;
>>> >> >> >>> >> >> >>>>>> at
>>> >> >> >>> >> >> >>>>>> the
end of command
>>> >> >> >>> >> >> >>>>>>
>>> >> >> >>> >> >> >>>>>> CREATE
TABLE pokes (foo INT, bar STRING);
>>> >> >> >>> >> >> >>>>>>
>>> >> >> >>> >> >> >>>>>> does'nt
work
>>> >> >> >>> >> >> >>>>>>
>>> >> >> >>> >> >> >>>>>>
>>> >> >> >>> >> >> >>>>>> On
Tue, Jun 5, 2012 at 8:34 PM, shashwat shriparv
>>> >> >> >>> >> >> >>>>>> <dwivedishashwat@gmail.com>
wrote:
>>> >> >> >>> >> >> >>>>>>>
>>> >> >> >>> >> >> >>>>>>>
inside configuration. all properties will be
>>> inside
>>> >> >> >>> >> >> >>>>>>>
the
>>> >> >> >>> >> >> >>>>>>>
configuration
>>> >> >> >>> >> >> >>>>>>>
tags
>>> >> >> >>> >> >> >>>>>>>
>>> >> >> >>> >> >> >>>>>>>
>>> >> >> >>> >> >> >>>>>>>
On Tue, Jun 5, 2012 at 11:53 PM, Babak Bastan
>>> >> >> >>> >> >> >>>>>>>
<babakbsn@gmail.com>
>>> >> >> >>> >> >> >>>>>>>
wrote:
>>> >> >> >>> >> >> >>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>
Thank you so much my friend your idee works
>>> fine(no
>>> >> >> >>> >> >> >>>>>>>>
error)
>>> >> >> >>> >> >> >>>>>>>>
you
>>> >> >> >>> >> >> >>>>>>>>
are
>>> >> >> >>> >> >> >>>>>>>>
the best :)
>>> >> >> >>> >> >> >>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>
On Tue, Jun 5, 2012 at 8:20 PM, Babak Bastan
>>> >> >> >>> >> >> >>>>>>>>
<babakbsn@gmail.com>
>>> >> >> >>> >> >> >>>>>>>>
wrote:
>>> >> >> >>> >> >> >>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>
It must be inside the
>>> >> >> >>> >> >> >>>>>>>>>
<configuration></configuration>
>>> >> >> >>> >> >> >>>>>>>>>
or
>>> >> >> >>> >> >> >>>>>>>>>
outside
>>> >> >> >>> >> >> >>>>>>>>>
this?
>>> >> >> >>> >> >> >>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>
On Tue, Jun 5, 2012 at 8:15 PM, shashwat
>>> shriparv
>>> >> >> >>> >> >> >>>>>>>>>
<dwivedishashwat@gmail.com> wrote:
>>> >> >> >>> >> >> >>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>
It will be inside hive/conf
>>> >> >> >>> >> >> >>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>
On Tue, Jun 5, 2012 at 11:43 PM, Babak Bastan
>>> >> >> >>> >> >> >>>>>>>>>>
<babakbsn@gmail.com>
>>> >> >> >>> >> >> >>>>>>>>>>
wrote:
>>> >> >> >>> >> >> >>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>
Thanks sShashwat, and where is this
>>> hive-site.xml
>>> >> >> >>> >> >> >>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>
On Tue, Jun 5, 2012 at 8:02 PM, shashwat
>>> shriparv
>>> >> >> >>> >> >> >>>>>>>>>>>
<dwivedishashwat@gmail.com> wrote:
>>> >> >> >>> >> >> >>>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>>
set
>>> >> >> >>> >> >> >>>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>>
hive.metastore.warehouse.dir in
>>> hive-site.xml
>>> >> >> >>> >> >> >>>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>>
<property>
>>> >> >> >>> >> >> >>>>>>>>>>>>
  <name>hive.metastore.local</name>
>>> >> >> >>> >> >> >>>>>>>>>>>>
  <value>true</value>
>>> >> >> >>> >> >> >>>>>>>>>>>>
</property>
>>> >> >> >>> >> >> >>>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>>
<name>hive.metastore.warehouse.dir</name>
>>> >> >> >>> >> >> >>>>>>>>>>>>
               <value>/home/<your
>>> >> >> >>> >> >> >>>>>>>>>>>>
username>/hivefolder</value>
>>> >> >> >>> >> >> >>>>>>>>>>>>
               <description>location of
>>> default
>>> >> >> >>> >> >> >>>>>>>>>>>>
database
>>> >> >> >>> >> >> >>>>>>>>>>>>
for
>>> >> >> >>> >> >> >>>>>>>>>>>>
the
>>> >> >> >>> >> >> >>>>>>>>>>>>
warehouse</description>
>>> >> >> >>> >> >> >>>>>>>>>>>>
       </property>
>>> >> >> >>> >> >> >>>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>>
On Tue, Jun 5, 2012 at 10:43 PM, Babak
>>> Bastan
>>> >> >> >>> >> >> >>>>>>>>>>>>
<babakbsn@gmail.com> wrote:
>>> >> >> >>> >> >> >>>>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>>>
Hello Experts ,
>>> >> >> >>> >> >> >>>>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>>>
I'm new in Hive .When try to create a test
>>> >> >> >>> >> >> >>>>>>>>>>>>>
Table
>>> >> >> >>> >> >> >>>>>>>>>>>>>
in
>>> >> >> >>> >> >> >>>>>>>>>>>>>
Hive
>>> >> >> >>> >> >> >>>>>>>>>>>>>
I
>>> >> >> >>> >> >> >>>>>>>>>>>>>
get
>>> >> >> >>> >> >> >>>>>>>>>>>>>
an error.I want to run this command:
>>> >> >> >>> >> >> >>>>>>>>>>>>>
CREATE TABLE Test (DateT STRING, Url
>>> STRING,
>>> >> >> >>> >> >> >>>>>>>>>>>>>
Content
>>> >> >> >>> >> >> >>>>>>>>>>>>>
STRING);
>>> >> >> >>> >> >> >>>>>>>>>>>>>
but this error occured:
>>> >> >> >>> >> >> >>>>>>>>>>>>>
FAILED: Error in metadata:
>>> >> >> >>> >> >> >>>>>>>>>>>>>
MetaException(message:Got
>>> >> >> >>> >> >> >>>>>>>>>>>>>
exception:
>>> >> >> >>> >> >> >>>>>>>>>>>>>
java.io.FileNotFoundException File
>>> >> >> >>> >> >> >>>>>>>>>>>>>
file:/user/hive/warehouse/test does not
>>> >> >> >>> >> >> >>>>>>>>>>>>>
exist.)
>>> >> >> >>> >> >> >>>>>>>>>>>>>
FAILED: Execution Error, return code 1 from
>>> >> >> >>> >> >> >>>>>>>>>>>>>
org.apache.hadoop.hive.ql.exec.DDLTask
>>> >> >> >>> >> >> >>>>>>>>>>>>>
How can I solve this Problem?
>>> >> >> >>> >> >> >>>>>>>>>>>>>
Thank you so much
>>> >> >> >>> >> >> >>>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>>
--
>>> >> >> >>> >> >> >>>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>>
∞
>>> >> >> >>> >> >> >>>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>>
Shashwat Shriparv
>>> >> >> >>> >> >> >>>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>
--
>>> >> >> >>> >> >> >>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>
∞
>>> >> >> >>> >> >> >>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>
Shashwat Shriparv
>>> >> >> >>> >> >> >>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>>
>>> >> >> >>> >> >> >>>>>>>>
>>> >> >> >>> >> >> >>>>>>>
>>> >> >> >>> >> >> >>>>>>>
>>> >> >> >>> >> >> >>>>>>>
>>> >> >> >>> >> >> >>>>>>>
--
>>> >> >> >>> >> >> >>>>>>>
>>> >> >> >>> >> >> >>>>>>>
>>> >> >> >>> >> >> >>>>>>>
∞
>>> >> >> >>> >> >> >>>>>>>
>>> >> >> >>> >> >> >>>>>>>
Shashwat Shriparv
>>> >> >> >>> >> >> >>>>>>>
>>> >> >> >>> >> >> >>>>>>>
>>> >> >> >>> >> >> >>>>>>
>>> >> >> >>> >> >> >>>>>
>>> >> >> >>> >> >> >>>>
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>> --
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>> ∞
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>> Shashwat Shriparv
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>>
>>> >> >> >>> >> >> >>
>>> >> >> >>> >> >
>>> >> >> >>> >> >
>>> >> >> >>> >
>>> >> >> >>> >
>>> >> >> >>
>>> >> >> >>
>>> >> >
>>> >> >
>>> >
>>> >
>>>
>>
>>
>
>
> --
>
>
> ∞
> Shashwat Shriparv
>
>
>

Mime
View raw message