hive-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From shashwat shriparv <dwivedishash...@gmail.com>
Subject Re: Error while Creating Table in Hive
Date Wed, 06 Jun 2012 20:02:47 GMT
whats the error babak ???

On Thu, Jun 7, 2012 at 1:25 AM, Babak Bastan <babakbsn@gmail.com> wrote:

> What the hell is that?I see no log folder there
>
>
> On Wed, Jun 6, 2012 at 9:41 PM, Mohammad Tariq <dontariq@gmail.com> wrote:
>
>> go to your HADOOP_HOME i.e your hadoop directory(that includes bin,
>> conf etc)..you can find logs directory there..
>>
>> Regards,
>>     Mohammad Tariq
>>
>>
>> On Thu, Jun 7, 2012 at 1:09 AM, Babak Bastan <babakbsn@gmail.com> wrote:
>> > hoe can I get my log mohammad?
>> >
>> >
>> > On Wed, Jun 6, 2012 at 9:36 PM, Mohammad Tariq <dontariq@gmail.com>
>> wrote:
>> >>
>> >> could you post your logs???that would help me in understanding the
>> >> problem properly.
>> >>
>> >> Regards,
>> >>     Mohammad Tariq
>> >>
>> >>
>> >> On Thu, Jun 7, 2012 at 1:02 AM, Babak Bastan <babakbsn@gmail.com>
>> wrote:
>> >> > Thank you very much mohamad for your attention.I followed the steps
>> but
>> >> > the
>> >> > error is the same as the last time.
>> >> > and there is my hosts file:
>> >> >
>> >> > 127.0.0.1       localhost
>> >> > #127.0.0.1      ubuntu.ubuntu-domain    ubuntu
>> >> >
>> >> >
>> >> > # The following lines are desirable for IPv6 capable hosts
>> >> >
>> >> > #::1     ip6-localhost ip6-loopback
>> >> > #fe00::0 ip6-localnet
>> >> > #ff00::0 ip6-mcastprefix
>> >> > #ff02::1 ip6-allnodes
>> >> > #ff02::2 ip6-allrouters
>> >> >
>> >> > but no effect :(
>> >> >
>> >> > On Wed, Jun 6, 2012 at 8:25 PM, Mohammad Tariq <dontariq@gmail.com>
>> >> > wrote:
>> >> >>
>> >> >> also change the permissions of these directories to 777.
>> >> >>
>> >> >> Regards,
>> >> >>     Mohammad Tariq
>> >> >>
>> >> >>
>> >> >> On Wed, Jun 6, 2012 at 11:54 PM, Mohammad Tariq <dontariq@gmail.com
>> >
>> >> >> wrote:
>> >> >> > create a directory "/home/username/hdfs" (or at some place
of your
>> >> >> > choice)..inside this hdfs directory create three sub directories
-
>> >> >> > name, data, and temp, then follow these steps :
>> >> >> >
>> >> >> > add following properties in your core-site.xml -
>> >> >> >
>> >> >> > <property>
>> >> >> >          <name>fs.default.name</name>
>> >> >> >          <value>hdfs://localhost:9000/</value>
>> >> >> >        </property>
>> >> >> >
>> >> >> >        <property>
>> >> >> >          <name>hadoop.tmp.dir</name>
>> >> >> >          <value>/home/mohammad/hdfs/temp</value>
>> >> >> >        </property>
>> >> >> >
>> >> >> > then add following two properties in your hdfs-site.xml -
>> >> >> >
>> >> >> > <property>
>> >> >> >                <name>dfs.replication</name>
>> >> >> >                <value>1</value>
>> >> >> >        </property>
>> >> >> >
>> >> >> >        <property>
>> >> >> >                <name>dfs.name.dir</name>
>> >> >> >                <value>/home/mohammad/hdfs/name</value>
>> >> >> >        </property>
>> >> >> >
>> >> >> >        <property>
>> >> >> >                <name>dfs.data.dir</name>
>> >> >> >                <value>/home/mohammad/hdfs/data</value>
>> >> >> >        </property>
>> >> >> >
>> >> >> > finally add this property in your mapred-site.xml -
>> >> >> >
>> >> >> >       <property>
>> >> >> >          <name>mapred.job.tracker</name>
>> >> >> >          <value>hdfs://localhost:9001</value>
>> >> >> >        </property>
>> >> >> >
>> >> >> > NOTE: you can give any name to these directories of your choice,
>> just
>> >> >> > keep in mind you have to give same names as values of
>> >> >> >           above specified properties in your configuration
files.
>> >> >> > (give full path of these directories, not just the name of
the
>> >> >> > directory)
>> >> >> >
>> >> >> > After this  follow the steps provided in the previous reply.
>> >> >> >
>> >> >> > Regards,
>> >> >> >     Mohammad Tariq
>> >> >> >
>> >> >> >
>> >> >> > On Wed, Jun 6, 2012 at 11:42 PM, Babak Bastan <babakbsn@gmail.com
>> >
>> >> >> > wrote:
>> >> >> >> thank's Mohammad
>> >> >> >>
>> >> >> >> with this command:
>> >> >> >>
>> >> >> >> babak@ubuntu:~/Downloads/hadoop/bin$ hadoop namenode -format
>> >> >> >>
>> >> >> >> this is my output:
>> >> >> >>
>> >> >> >> 12/06/06 20:05:20 INFO namenode.NameNode: STARTUP_MSG:
>> >> >> >> /************************************************************
>> >> >> >> STARTUP_MSG: Starting NameNode
>> >> >> >> STARTUP_MSG:   host = ubuntu/127.0.1.1
>> >> >> >> STARTUP_MSG:   args = [-format]
>> >> >> >> STARTUP_MSG:   version = 0.20.2
>> >> >> >> STARTUP_MSG:   build =
>> >> >> >>
>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20
>> >> >> >> -r
>> >> >> >> 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC
2010
>> >> >> >> ************************************************************/
>> >> >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem:
>> >> >> >>
>> >> >> >>
>> fsOwner=babak,babak,adm,dialout,cdrom,plugdev,lpadmin,admin,sambashare
>> >> >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem:
>> supergroup=supergroup
>> >> >> >> 12/06/06 20:05:20 INFO namenode.FSNamesystem:
>> >> >> >> isPermissionEnabled=true
>> >> >> >> 12/06/06 20:05:20 INFO common.Storage: Image file of size
95
>> saved
>> >> >> >> in 0
>> >> >> >> seconds.
>> >> >> >> 12/06/06 20:05:20 INFO common.Storage: Storage directory
>> >> >> >> /tmp/hadoop-babak/dfs/name has been successfully formatted.
>> >> >> >> 12/06/06 20:05:20 INFO namenode.NameNode: SHUTDOWN_MSG:
>> >> >> >> /************************************************************
>> >> >> >> SHUTDOWN_MSG: Shutting down NameNode at ubuntu/127.0.1.1
>> >> >> >> ************************************************************/
>> >> >> >>
>> >> >> >> by this command:
>> >> >> >>
>> >> >> >> babak@ubuntu:~/Downloads/hadoop/bin$ start-dfs.sh
>> >> >> >>
>> >> >> >> this is the out put
>> >> >> >>
>> >> >> >> mkdir: kann Verzeichnis
>> „/home/babak/Downloads/hadoop/bin/../logs“
>> >> >> >> nicht
>> >> >> >> anlegen: Keine Berechtigung
>> >> >> >>
>> >> >> >> this out put(it's in german and it means no right to make
this
>> >> >> >> folder)
>> >> >> >>
>> >> >> >>
>> >> >> >> On Wed, Jun 6, 2012 at 7:59 PM, Mohammad Tariq <
>> dontariq@gmail.com>
>> >> >> >> wrote:
>> >> >> >>>
>> >> >> >>> once we are done with the configuration, we need to
format the
>> file
>> >> >> >>> system..use this command to do that-
>> >> >> >>> bin/hadoop namenode -format
>> >> >> >>>
>> >> >> >>> after this, hadoop daemon processes should be started
using
>> >> >> >>> following
>> >> >> >>> commands -
>> >> >> >>> bin/start-dfs.sh (it'll start NN & DN)
>> >> >> >>> bin/start-mapred.sh (it'll start JT & TT)
>> >> >> >>>
>> >> >> >>> after this use jps to check if everything is alright
or point
>> your
>> >> >> >>> browser to localhost:50070..if you further find any
problem
>> provide
>> >> >> >>> us
>> >> >> >>> with the error logs..:)
>> >> >> >>>
>> >> >> >>> Regards,
>> >> >> >>>     Mohammad Tariq
>> >> >> >>>
>> >> >> >>>
>> >> >> >>> On Wed, Jun 6, 2012 at 11:22 PM, Babak Bastan <
>> babakbsn@gmail.com>
>> >> >> >>> wrote:
>> >> >> >>> > were you able to format hdfs properly???
>> >> >> >>> > I did'nt get your question,Do you mean HADOOP_HOME?
or where
>> did
>> >> >> >>> > I
>> >> >> >>> > install
>> >> >> >>> > Hadoop?
>> >> >> >>> >
>> >> >> >>> > On Wed, Jun 6, 2012 at 7:49 PM, Mohammad Tariq
>> >> >> >>> > <dontariq@gmail.com>
>> >> >> >>> > wrote:
>> >> >> >>> >>
>> >> >> >>> >> if you are getting only this, it means your
hadoop is not
>> >> >> >>> >> running..were you able to format hdfs properly???
>> >> >> >>> >>
>> >> >> >>> >> Regards,
>> >> >> >>> >>     Mohammad Tariq
>> >> >> >>> >>
>> >> >> >>> >>
>> >> >> >>> >> On Wed, Jun 6, 2012 at 11:17 PM, Babak Bastan
>> >> >> >>> >> <babakbsn@gmail.com>
>> >> >> >>> >> wrote:
>> >> >> >>> >> > Hi MohammadmI irun jps in my shel I
can see this result:
>> >> >> >>> >> > 2213 Jps
>> >> >> >>> >> >
>> >> >> >>> >> >
>> >> >> >>> >> > On Wed, Jun 6, 2012 at 7:44 PM, Mohammad
Tariq
>> >> >> >>> >> > <dontariq@gmail.com>
>> >> >> >>> >> > wrote:
>> >> >> >>> >> >>
>> >> >> >>> >> >> you can also use "jps" command at
your shell to see
>> whether
>> >> >> >>> >> >> Hadoop
>> >> >> >>> >> >> processes are running or not.
>> >> >> >>> >> >>
>> >> >> >>> >> >> Regards,
>> >> >> >>> >> >>     Mohammad Tariq
>> >> >> >>> >> >>
>> >> >> >>> >> >>
>> >> >> >>> >> >> On Wed, Jun 6, 2012 at 11:12 PM,
Mohammad Tariq
>> >> >> >>> >> >> <dontariq@gmail.com>
>> >> >> >>> >> >> wrote:
>> >> >> >>> >> >> > Hi Babak,
>> >> >> >>> >> >> >
>> >> >> >>> >> >> >  You have to type it in you
web browser..Hadoop
>> provides us
>> >> >> >>> >> >> > a
>> >> >> >>> >> >> > web
>> >> >> >>> >> >> > GUI
>> >> >> >>> >> >> > that not only allows us to
browse through the file
>> system,
>> >> >> >>> >> >> > but
>> >> >> >>> >> >> > to
>> >> >> >>> >> >> > download the files as well..Apart
from that it also
>> >> >> >>> >> >> > provides a
>> >> >> >>> >> >> > web
>> >> >> >>> >> >> > GUI
>> >> >> >>> >> >> > that can be used to see the
status of Jobtracker and
>> >> >> >>> >> >> > Tasktracker..When
>> >> >> >>> >> >> > you run a Hive or Pig job or
a Mapreduce job, you can
>> point
>> >> >> >>> >> >> > your
>> >> >> >>> >> >> > browser to http://localhost:50030
to see the status and
>> >> >> >>> >> >> > logs
>> >> >> >>> >> >> > of
>> >> >> >>> >> >> > your
>> >> >> >>> >> >> > job.
>> >> >> >>> >> >> >
>> >> >> >>> >> >> > Regards,
>> >> >> >>> >> >> >     Mohammad Tariq
>> >> >> >>> >> >> >
>> >> >> >>> >> >> >
>> >> >> >>> >> >> > On Wed, Jun 6, 2012 at 8:28
PM, Babak Bastan
>> >> >> >>> >> >> > <babakbsn@gmail.com>
>> >> >> >>> >> >> > wrote:
>> >> >> >>> >> >> >> Thank you shashwat for
the answer,
>> >> >> >>> >> >> >> where should I type http://localhost:50070?
>> >> >> >>> >> >> >> I typed here: hive>http://localhost:50070
but nothing
>> as
>> >> >> >>> >> >> >> result
>> >> >> >>> >> >> >>
>> >> >> >>> >> >> >>
>> >> >> >>> >> >> >> On Wed, Jun 6, 2012 at
3:32 PM, shashwat shriparv
>> >> >> >>> >> >> >> <dwivedishashwat@gmail.com>
wrote:
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>> first type http://localhost:50070
whether this is
>> opening
>> >> >> >>> >> >> >>> or
>> >> >> >>> >> >> >>> not
>> >> >> >>> >> >> >>> and
>> >> >> >>> >> >> >>> check
>> >> >> >>> >> >> >>> how many nodes are
available, check some of the hadoop
>> >> >> >>> >> >> >>> shell
>> >> >> >>> >> >> >>> commands
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>> from
>> http://hadoop.apache.org/common/docs/r0.18.3/hdfs_shell.html
>> >> >> >>> >> >> >>> run
>> >> >> >>> >> >> >>> example mapreduce task
on hadoop take example from
>> here
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>> :
>> http://www.michael-noll.com/blog/2011/04/09/benchmarking-and-stress-testing-an-hadoop-cluster-with-terasort-testdfsio-nnbench-mrbench/
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>> if all the above you
can do sucessfully means hadoop
>> is
>> >> >> >>> >> >> >>> configured
>> >> >> >>> >> >> >>> correctly
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>> Regards
>> >> >> >>> >> >> >>> Shashwat
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>> On Wed, Jun 6, 2012
at 1:30 AM, Babak Bastan
>> >> >> >>> >> >> >>> <babakbsn@gmail.com>
>> >> >> >>> >> >> >>> wrote:
>> >> >> >>> >> >> >>>>
>> >> >> >>> >> >> >>>> no I'm not working
on CDH.Is there a way to test if
>> my
>> >> >> >>> >> >> >>>> Hadoop
>> >> >> >>> >> >> >>>> works
>> >> >> >>> >> >> >>>> fine
>> >> >> >>> >> >> >>>> or not?
>> >> >> >>> >> >> >>>>
>> >> >> >>> >> >> >>>>
>> >> >> >>> >> >> >>>> On Tue, Jun 5,
2012 at 9:55 PM, Bejoy KS
>> >> >> >>> >> >> >>>> <bejoy_ks@yahoo.com>
>> >> >> >>> >> >> >>>> wrote:
>> >> >> >>> >> >> >>>>>
>> >> >> >>> >> >> >>>>> Hi Babak
>> >> >> >>> >> >> >>>>>
>> >> >> >>> >> >> >>>>> You gotta follow
those instructions in the apace
>> site
>> >> >> >>> >> >> >>>>> to
>> >> >> >>> >> >> >>>>> set
>> >> >> >>> >> >> >>>>> up
>> >> >> >>> >> >> >>>>> hadoop
>> >> >> >>> >> >> >>>>> from scratch
and ensure that hdfs is working first.
>> You
>> >> >> >>> >> >> >>>>> should
>> >> >> >>> >> >> >>>>> be
>> >> >> >>> >> >> >>>>> able to
>> >> >> >>> >> >> >>>>> read and write
files to hdfs before you do your next
>> >> >> >>> >> >> >>>>> steps.
>> >> >> >>> >> >> >>>>>
>> >> >> >>> >> >> >>>>> Are you on
CDH or apache distribution of hadoop? If
>> it
>> >> >> >>> >> >> >>>>> is
>> >> >> >>> >> >> >>>>> CDH
>> >> >> >>> >> >> >>>>> there
>> >> >> >>> >> >> >>>>> are
>> >> >> >>> >> >> >>>>> detailed instructions
on Cloudera web site.
>> >> >> >>> >> >> >>>>>
>> >> >> >>> >> >> >>>>> Regards
>> >> >> >>> >> >> >>>>> Bejoy KS
>> >> >> >>> >> >> >>>>>
>> >> >> >>> >> >> >>>>> Sent from handheld,
please excuse typos.
>> >> >> >>> >> >> >>>>> ________________________________
>> >> >> >>> >> >> >>>>> From: Babak
Bastan <babakbsn@gmail.com>
>> >> >> >>> >> >> >>>>> Date: Tue,
5 Jun 2012 21:30:22 +0200
>> >> >> >>> >> >> >>>>> To: <user@hive.apache.org>
>> >> >> >>> >> >> >>>>> ReplyTo: user@hive.apache.org
>> >> >> >>> >> >> >>>>> Subject: Re:
Error while Creating Table in Hive
>> >> >> >>> >> >> >>>>>
>> >> >> >>> >> >> >>>>> @Bejoy: I set
the fs.default.name in the
>> core-site.xml
>> >> >> >>> >> >> >>>>> and
>> >> >> >>> >> >> >>>>> I
>> >> >> >>> >> >> >>>>> did
>> >> >> >>> >> >> >>>>> all
>> >> >> >>> >> >> >>>>> of
>> >> >> >>> >> >> >>>>> thing that
was mentioned in the reference but no
>> effect
>> >> >> >>> >> >> >>>>>
>> >> >> >>> >> >> >>>>> On Tue, Jun
5, 2012 at 8:43 PM, Babak Bastan
>> >> >> >>> >> >> >>>>> <babakbsn@gmail.com>
>> >> >> >>> >> >> >>>>> wrote:
>> >> >> >>> >> >> >>>>>>
>> >> >> >>> >> >> >>>>>> Ok sorry
but that was my Mistake .I thought it
>> works
>> >> >> >>> >> >> >>>>>> but
>> >> >> >>> >> >> >>>>>> no.
>> >> >> >>> >> >> >>>>>> I wrote
the command without ; and then I think It
>> >> >> >>> >> >> >>>>>> works
>> >> >> >>> >> >> >>>>>> but
>> >> >> >>> >> >> >>>>>> with
>> >> >> >>> >> >> >>>>>> ;
>> >> >> >>> >> >> >>>>>> at
>> >> >> >>> >> >> >>>>>> the end
of command
>> >> >> >>> >> >> >>>>>>
>> >> >> >>> >> >> >>>>>> CREATE
TABLE pokes (foo INT, bar STRING);
>> >> >> >>> >> >> >>>>>>
>> >> >> >>> >> >> >>>>>> does'nt
work
>> >> >> >>> >> >> >>>>>>
>> >> >> >>> >> >> >>>>>>
>> >> >> >>> >> >> >>>>>> On Tue,
Jun 5, 2012 at 8:34 PM, shashwat shriparv
>> >> >> >>> >> >> >>>>>> <dwivedishashwat@gmail.com>
wrote:
>> >> >> >>> >> >> >>>>>>>
>> >> >> >>> >> >> >>>>>>> inside
configuration. all properties will be
>> inside
>> >> >> >>> >> >> >>>>>>> the
>> >> >> >>> >> >> >>>>>>> configuration
>> >> >> >>> >> >> >>>>>>> tags
>> >> >> >>> >> >> >>>>>>>
>> >> >> >>> >> >> >>>>>>>
>> >> >> >>> >> >> >>>>>>> On
Tue, Jun 5, 2012 at 11:53 PM, Babak Bastan
>> >> >> >>> >> >> >>>>>>> <babakbsn@gmail.com>
>> >> >> >>> >> >> >>>>>>> wrote:
>> >> >> >>> >> >> >>>>>>>>
>> >> >> >>> >> >> >>>>>>>>
Thank you so much my friend your idee works
>> fine(no
>> >> >> >>> >> >> >>>>>>>>
error)
>> >> >> >>> >> >> >>>>>>>>
you
>> >> >> >>> >> >> >>>>>>>>
are
>> >> >> >>> >> >> >>>>>>>>
the best :)
>> >> >> >>> >> >> >>>>>>>>
>> >> >> >>> >> >> >>>>>>>>
>> >> >> >>> >> >> >>>>>>>>
On Tue, Jun 5, 2012 at 8:20 PM, Babak Bastan
>> >> >> >>> >> >> >>>>>>>>
<babakbsn@gmail.com>
>> >> >> >>> >> >> >>>>>>>>
wrote:
>> >> >> >>> >> >> >>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>
It must be inside the
>> >> >> >>> >> >> >>>>>>>>>
<configuration></configuration>
>> >> >> >>> >> >> >>>>>>>>>
or
>> >> >> >>> >> >> >>>>>>>>>
outside
>> >> >> >>> >> >> >>>>>>>>>
this?
>> >> >> >>> >> >> >>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>
On Tue, Jun 5, 2012 at 8:15 PM, shashwat
>> shriparv
>> >> >> >>> >> >> >>>>>>>>>
<dwivedishashwat@gmail.com> wrote:
>> >> >> >>> >> >> >>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>
It will be inside hive/conf
>> >> >> >>> >> >> >>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>
On Tue, Jun 5, 2012 at 11:43 PM, Babak Bastan
>> >> >> >>> >> >> >>>>>>>>>>
<babakbsn@gmail.com>
>> >> >> >>> >> >> >>>>>>>>>>
wrote:
>> >> >> >>> >> >> >>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>
Thanks sShashwat, and where is this
>> hive-site.xml
>> >> >> >>> >> >> >>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>
On Tue, Jun 5, 2012 at 8:02 PM, shashwat
>> shriparv
>> >> >> >>> >> >> >>>>>>>>>>>
<dwivedishashwat@gmail.com> wrote:
>> >> >> >>> >> >> >>>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>>
set
>> >> >> >>> >> >> >>>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>>
hive.metastore.warehouse.dir in hive-site.xml
>> >> >> >>> >> >> >>>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>>
<property>
>> >> >> >>> >> >> >>>>>>>>>>>>
  <name>hive.metastore.local</name>
>> >> >> >>> >> >> >>>>>>>>>>>>
  <value>true</value>
>> >> >> >>> >> >> >>>>>>>>>>>>
</property>
>> >> >> >>> >> >> >>>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>>
<name>hive.metastore.warehouse.dir</name>
>> >> >> >>> >> >> >>>>>>>>>>>>
               <value>/home/<your
>> >> >> >>> >> >> >>>>>>>>>>>>
username>/hivefolder</value>
>> >> >> >>> >> >> >>>>>>>>>>>>
               <description>location of
>> default
>> >> >> >>> >> >> >>>>>>>>>>>>
database
>> >> >> >>> >> >> >>>>>>>>>>>>
for
>> >> >> >>> >> >> >>>>>>>>>>>>
the
>> >> >> >>> >> >> >>>>>>>>>>>>
warehouse</description>
>> >> >> >>> >> >> >>>>>>>>>>>>
       </property>
>> >> >> >>> >> >> >>>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>>
On Tue, Jun 5, 2012 at 10:43 PM, Babak Bastan
>> >> >> >>> >> >> >>>>>>>>>>>>
<babakbsn@gmail.com> wrote:
>> >> >> >>> >> >> >>>>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>>>
Hello Experts ,
>> >> >> >>> >> >> >>>>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>>>
I'm new in Hive .When try to create a test
>> >> >> >>> >> >> >>>>>>>>>>>>>
Table
>> >> >> >>> >> >> >>>>>>>>>>>>>
in
>> >> >> >>> >> >> >>>>>>>>>>>>>
Hive
>> >> >> >>> >> >> >>>>>>>>>>>>>
I
>> >> >> >>> >> >> >>>>>>>>>>>>>
get
>> >> >> >>> >> >> >>>>>>>>>>>>>
an error.I want to run this command:
>> >> >> >>> >> >> >>>>>>>>>>>>>
CREATE TABLE Test (DateT STRING, Url STRING,
>> >> >> >>> >> >> >>>>>>>>>>>>>
Content
>> >> >> >>> >> >> >>>>>>>>>>>>>
STRING);
>> >> >> >>> >> >> >>>>>>>>>>>>>
but this error occured:
>> >> >> >>> >> >> >>>>>>>>>>>>>
FAILED: Error in metadata:
>> >> >> >>> >> >> >>>>>>>>>>>>>
MetaException(message:Got
>> >> >> >>> >> >> >>>>>>>>>>>>>
exception:
>> >> >> >>> >> >> >>>>>>>>>>>>>
java.io.FileNotFoundException File
>> >> >> >>> >> >> >>>>>>>>>>>>>
file:/user/hive/warehouse/test does not
>> >> >> >>> >> >> >>>>>>>>>>>>>
exist.)
>> >> >> >>> >> >> >>>>>>>>>>>>>
FAILED: Execution Error, return code 1 from
>> >> >> >>> >> >> >>>>>>>>>>>>>
org.apache.hadoop.hive.ql.exec.DDLTask
>> >> >> >>> >> >> >>>>>>>>>>>>>
How can I solve this Problem?
>> >> >> >>> >> >> >>>>>>>>>>>>>
Thank you so much
>> >> >> >>> >> >> >>>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>>
--
>> >> >> >>> >> >> >>>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>>
∞
>> >> >> >>> >> >> >>>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>>
Shashwat Shriparv
>> >> >> >>> >> >> >>>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>
--
>> >> >> >>> >> >> >>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>
∞
>> >> >> >>> >> >> >>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>
Shashwat Shriparv
>> >> >> >>> >> >> >>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>>
>> >> >> >>> >> >> >>>>>>>>
>> >> >> >>> >> >> >>>>>>>
>> >> >> >>> >> >> >>>>>>>
>> >> >> >>> >> >> >>>>>>>
>> >> >> >>> >> >> >>>>>>> --
>> >> >> >>> >> >> >>>>>>>
>> >> >> >>> >> >> >>>>>>>
>> >> >> >>> >> >> >>>>>>> ∞
>> >> >> >>> >> >> >>>>>>>
>> >> >> >>> >> >> >>>>>>> Shashwat
Shriparv
>> >> >> >>> >> >> >>>>>>>
>> >> >> >>> >> >> >>>>>>>
>> >> >> >>> >> >> >>>>>>
>> >> >> >>> >> >> >>>>>
>> >> >> >>> >> >> >>>>
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>> --
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>> ∞
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>> Shashwat Shriparv
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>>
>> >> >> >>> >> >> >>
>> >> >> >>> >> >
>> >> >> >>> >> >
>> >> >> >>> >
>> >> >> >>> >
>> >> >> >>
>> >> >> >>
>> >> >
>> >> >
>> >
>> >
>>
>
>


-- 


∞
Shashwat Shriparv

Mime
View raw message