I thought it was good to go.
Hey, have you tried maybe just doing a simple load test? I mean just loading
a file into grunt with the LOAD command, and then doing a DUMP on it. So
after that, we could see if there is actually something wrong with your
installation.


Renato M.

2010/9/27 Ngô Văn Vĩ <ngovi.se....@gmail.com>

> "192.168.230.130" is IP of my machine
> @JeffZhang: can you explain clearly?
> Thanks
>
> On Tue, Sep 28, 2010 at 8:39 AM, Jeff Zhang <zjf...@gmail.com> wrote:
>
> > It seems you have connected to the right hadoop when you start pig
> > grunt. But connect to the wrong hadoop when you run pig script.
> > Try to search whether there's other configuration files that mess up
> > with your default configuration. And what is machine "192.168.230.130"
> > ?
> >
> >
> > On Tue, Sep 28, 2010 at 9:23 AM, Ngô Văn Vĩ <ngovi.se....@gmail.com>
> > wrote:
> > > have you help me?
> > > i have configuration
> > > *-  bin/pig*
> > > export JAVA_HOME=/home/ngovi/jdk1.6.0_21
> > > export PIG_INSTALL=/home/ngovi/pig-0.7.0
> > > export PATH=$PATH:$PIG_INSTALL/bin
> > > export PIG_HADOOP_VERSION=0.20.2
> > > export PIG_CLASSPATH=/home/ngovi/hadoop-0.20.2/conf/
> > > ....
> > > *- conf/pig.properties*
> > > fs.default.name=hdfs://localhost:9000/
> > > mapred.job.tracker=localhost:9001
> > > # log4jconf log4j configuration file
> > > i run pig that have error
> > >
> > > *- in hadoop-0.20.2/conf*
> > > *core-site.xml*
> > > <configuration>
> > > <property>
> > > <name>fs.default.name</name>
> > > <value>hdfs://localhost:9000</value>
> > > <description>
> > > the name of the default file system
> > > </description>
> > > </property>
> > > </configuration>
> > > *hdfs-site.xml*
> > > <configuration>
> > > <property>
> > > <name>dfs.replication</name>
> > > <value>1</value>
> > > <description>Default block replication </description>
> > > </property>
> > > </configuration>
> > >
> > > *mapred-site.xml*
> > >
> > > <configuration>
> > > <property>
> > > <name>mapred.job.tracker</name>
> > > <value>localhost:9001</value>
> > > <description>
> > > the host and port that the mapreduce job tracker run at
> > > </description>
> > > </property>
> > > </configuration>
> > >
> > > I run pig that have error??
> > > *ng...@master:~/pig-0.7.0$ bin/pig -x mapreduce
> > > 10/09/27 18:16:29 INFO pig.Main: Logging error messages to:
> > > /home/ngovi/pig-0.7.0/pig_1285636589590.log
> > > 2010-09-27 18:16:30,029 [main] INFO
> > > org.apache.pig.backend.hadoop.executionengine.HExecutionEngine -
> > Connecting
> > > to hadoop file system at: hdfs://localhost:9000/
> > > 2010-09-27 18:16:30,347 [main] INFO
> > > org.apache.pig.backend.hadoop.executionengine.HExecutionEngine -
> > Connecting
> > > to map-reduce job tracker at: localhost:9001
> > > grunt> *
> > >
> > >
> > > thanks all
> > >
> > > On Mon, Sep 27, 2010 at 1:14 PM, Alan Gates <ga...@yahoo-inc.com>
> wrote:
> > >
> > >> Pig is failing to connect to your namenode.  Is the address Pig is
> > trying
> > >> to use (hdfs://master:54310/) correct?  Can you connect using that
> > string
> > >> from the same machine using bin/hadoop?
> > >>
> > >> Alan.
> > >>
> > >>
> > >> On Sep 27, 2010, at 8:45 AM, Ngô Văn Vĩ wrote:
> > >>
> > >>  I run Pig at Hadoop Mode
> > >>> (Pig-0.7.0 and hadoop-0.20.2)
> > >>> have error?
> > >>> ng...@master:~/pig-0.7.0$ bin/pig
> > >>> 10/09/27 08:39:40 INFO pig.Main: Logging error messages to:
> > >>> /home/ngovi/pig-0.7.0/pig_1285601980268.log
> > >>> 2010-09-27 08:39:40,538 [main] INFO
> > >>> org.apache.pig.backend.hadoop.executionengine.HExecutionEngine -
> > >>> Connecting
> > >>> to hadoop file system at: hdfs://master:54310/
> > >>> 2010-09-27 08:39:41,760 [main] INFO  org.apache.hadoop.ipc.Client -
> > >>> Retrying
> > >>> connect to server: master/192.168.230.130:54310. Already tried 0
> > time(s).
> > >>> 2010-09-27 08:39:42,762 [main] INFO  org.apache.hadoop.ipc.Client -
> > >>> Retrying
> > >>> connect to server: master/192.168.230.130:54310. Already tried 1
> > time(s).
> > >>> 2010-09-27 08:39:43,763 [main] INFO  org.apache.hadoop.ipc.Client -
> > >>> Retrying
> > >>> connect to server: master/192.168.230.130:54310. Already tried 2
> > time(s).
> > >>> 2010-09-27 08:39:44,765 [main] INFO  org.apache.hadoop.ipc.Client -
> > >>> Retrying
> > >>> connect to server: master/192.168.230.130:54310. Already tried 3
> > time(s).
> > >>> 2010-09-27 08:39:45,766 [main] INFO  org.apache.hadoop.ipc.Client -
> > >>> Retrying
> > >>> connect to server: master/192.168.230.130:54310. Already tried 4
> > time(s).
> > >>> 2010-09-27 08:39:46,767 [main] INFO  org.apache.hadoop.ipc.Client -
> > >>> Retrying
> > >>> connect to server: master/192.168.230.130:54310. Already tried 5
> > time(s).
> > >>> 2010-09-27 08:39:47,768 [main] INFO  org.apache.hadoop.ipc.Client -
> > >>> Retrying
> > >>> connect to server: master/192.168.230.130:54310. Already tried 6
> > time(s).
> > >>> 2010-09-27 08:39:48,769 [main] INFO  org.apache.hadoop.ipc.Client -
> > >>> Retrying
> > >>> connect to server: master/192.168.230.130:54310. Already tried 7
> > time(s).
> > >>> 2010-09-27 08:39:49,770 [main] INFO  org.apache.hadoop.ipc.Client -
> > >>> Retrying
> > >>> connect to server: master/192.168.230.130:54310. Already tried 8
> > time(s).
> > >>> 2010-09-27 08:39:50,771 [main] INFO  org.apache.hadoop.ipc.Client -
> > >>> Retrying
> > >>> connect to server: master/192.168.230.130:54310. Already tried 9
> > time(s).
> > >>> 2010-09-27 08:39:50,780 [main] ERROR org.apache.pig.Main - ERROR
> 2999:
> > >>> Unexpected internal error. Failed to create DataStorage
> > >>>
> > >>> Help me??
> > >>> Thanks
> > >>> --
> > >>> Ngô Văn Vĩ
> > >>> Công Nghệ Phần Mềm
> > >>> Phone: 01695893851
> > >>>
> > >>
> > >>
> > >
> > >
> > > --
> > > Ngô Văn Vĩ
> > > Công Nghệ Phần Mềm
> > > Phone: 01695893851
> > >
> >
> >
> >
> > --
> > Best Regards
> >
> > Jeff Zhang
> >
>
>
>
> --
> Ngô Văn Vĩ
> Công Nghệ Phần Mềm
> Phone: 01695893851
>

Reply via email to