hadoop-pig-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Ngô Văn Vĩ <ngovi.se....@gmail.com>
Subject Re: help : error run pig
Date Tue, 28 Sep 2010 02:22:31 GMT
"192.168.230.130" is IP of my machine
@JeffZhang: can you explain clearly?
Thanks

On Tue, Sep 28, 2010 at 8:39 AM, Jeff Zhang <zjffdu@gmail.com> wrote:

> It seems you have connected to the right hadoop when you start pig
> grunt. But connect to the wrong hadoop when you run pig script.
> Try to search whether there's other configuration files that mess up
> with your default configuration. And what is machine "192.168.230.130"
> ?
>
>
> On Tue, Sep 28, 2010 at 9:23 AM, Ngô Văn Vĩ <ngovi.se.fit@gmail.com>
> wrote:
> > have you help me?
> > i have configuration
> > *-  bin/pig*
> > export JAVA_HOME=/home/ngovi/jdk1.6.0_21
> > export PIG_INSTALL=/home/ngovi/pig-0.7.0
> > export PATH=$PATH:$PIG_INSTALL/bin
> > export PIG_HADOOP_VERSION=0.20.2
> > export PIG_CLASSPATH=/home/ngovi/hadoop-0.20.2/conf/
> > ....
> > *- conf/pig.properties*
> > fs.default.name=hdfs://localhost:9000/
> > mapred.job.tracker=localhost:9001
> > # log4jconf log4j configuration file
> > i run pig that have error
> >
> > *- in hadoop-0.20.2/conf*
> > *core-site.xml*
> > <configuration>
> > <property>
> > <name>fs.default.name</name>
> > <value>hdfs://localhost:9000</value>
> > <description>
> > the name of the default file system
> > </description>
> > </property>
> > </configuration>
> > *hdfs-site.xml*
> > <configuration>
> > <property>
> > <name>dfs.replication</name>
> > <value>1</value>
> > <description>Default block replication </description>
> > </property>
> > </configuration>
> >
> > *mapred-site.xml*
> >
> > <configuration>
> > <property>
> > <name>mapred.job.tracker</name>
> > <value>localhost:9001</value>
> > <description>
> > the host and port that the mapreduce job tracker run at
> > </description>
> > </property>
> > </configuration>
> >
> > I run pig that have error??
> > *ngovi@master:~/pig-0.7.0$ bin/pig -x mapreduce
> > 10/09/27 18:16:29 INFO pig.Main: Logging error messages to:
> > /home/ngovi/pig-0.7.0/pig_1285636589590.log
> > 2010-09-27 18:16:30,029 [main] INFO
> > org.apache.pig.backend.hadoop.executionengine.HExecutionEngine -
> Connecting
> > to hadoop file system at: hdfs://localhost:9000/
> > 2010-09-27 18:16:30,347 [main] INFO
> > org.apache.pig.backend.hadoop.executionengine.HExecutionEngine -
> Connecting
> > to map-reduce job tracker at: localhost:9001
> > grunt> *
> >
> >
> > thanks all
> >
> > On Mon, Sep 27, 2010 at 1:14 PM, Alan Gates <gates@yahoo-inc.com> wrote:
> >
> >> Pig is failing to connect to your namenode.  Is the address Pig is
> trying
> >> to use (hdfs://master:54310/) correct?  Can you connect using that
> string
> >> from the same machine using bin/hadoop?
> >>
> >> Alan.
> >>
> >>
> >> On Sep 27, 2010, at 8:45 AM, Ngô Văn Vĩ wrote:
> >>
> >>  I run Pig at Hadoop Mode
> >>> (Pig-0.7.0 and hadoop-0.20.2)
> >>> have error?
> >>> ngovi@master:~/pig-0.7.0$ bin/pig
> >>> 10/09/27 08:39:40 INFO pig.Main: Logging error messages to:
> >>> /home/ngovi/pig-0.7.0/pig_1285601980268.log
> >>> 2010-09-27 08:39:40,538 [main] INFO
> >>> org.apache.pig.backend.hadoop.executionengine.HExecutionEngine -
> >>> Connecting
> >>> to hadoop file system at: hdfs://master:54310/
> >>> 2010-09-27 08:39:41,760 [main] INFO  org.apache.hadoop.ipc.Client -
> >>> Retrying
> >>> connect to server: master/192.168.230.130:54310. Already tried 0
> time(s).
> >>> 2010-09-27 08:39:42,762 [main] INFO  org.apache.hadoop.ipc.Client -
> >>> Retrying
> >>> connect to server: master/192.168.230.130:54310. Already tried 1
> time(s).
> >>> 2010-09-27 08:39:43,763 [main] INFO  org.apache.hadoop.ipc.Client -
> >>> Retrying
> >>> connect to server: master/192.168.230.130:54310. Already tried 2
> time(s).
> >>> 2010-09-27 08:39:44,765 [main] INFO  org.apache.hadoop.ipc.Client -
> >>> Retrying
> >>> connect to server: master/192.168.230.130:54310. Already tried 3
> time(s).
> >>> 2010-09-27 08:39:45,766 [main] INFO  org.apache.hadoop.ipc.Client -
> >>> Retrying
> >>> connect to server: master/192.168.230.130:54310. Already tried 4
> time(s).
> >>> 2010-09-27 08:39:46,767 [main] INFO  org.apache.hadoop.ipc.Client -
> >>> Retrying
> >>> connect to server: master/192.168.230.130:54310. Already tried 5
> time(s).
> >>> 2010-09-27 08:39:47,768 [main] INFO  org.apache.hadoop.ipc.Client -
> >>> Retrying
> >>> connect to server: master/192.168.230.130:54310. Already tried 6
> time(s).
> >>> 2010-09-27 08:39:48,769 [main] INFO  org.apache.hadoop.ipc.Client -
> >>> Retrying
> >>> connect to server: master/192.168.230.130:54310. Already tried 7
> time(s).
> >>> 2010-09-27 08:39:49,770 [main] INFO  org.apache.hadoop.ipc.Client -
> >>> Retrying
> >>> connect to server: master/192.168.230.130:54310. Already tried 8
> time(s).
> >>> 2010-09-27 08:39:50,771 [main] INFO  org.apache.hadoop.ipc.Client -
> >>> Retrying
> >>> connect to server: master/192.168.230.130:54310. Already tried 9
> time(s).
> >>> 2010-09-27 08:39:50,780 [main] ERROR org.apache.pig.Main - ERROR 2999:
> >>> Unexpected internal error. Failed to create DataStorage
> >>>
> >>> Help me??
> >>> Thanks
> >>> --
> >>> Ngô Văn Vĩ
> >>> Công Nghệ Phần Mềm
> >>> Phone: 01695893851
> >>>
> >>
> >>
> >
> >
> > --
> > Ngô Văn Vĩ
> > Công Nghệ Phần Mềm
> > Phone: 01695893851
> >
>
>
>
> --
> Best Regards
>
> Jeff Zhang
>



-- 
Ngô Văn Vĩ
Công Nghệ Phần Mềm
Phone: 01695893851

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message