I thought it was good to go. Hey, have you tried maybe just doing a simple load test? I mean just loading a file into grunt with the LOAD command, and then doing a DUMP on it. So after that, we could see if there is actually something wrong with your installation.
Renato M. 2010/9/27 Ngô Văn Vĩ <ngovi.se....@gmail.com> > "192.168.230.130" is IP of my machine > @JeffZhang: can you explain clearly? > Thanks > > On Tue, Sep 28, 2010 at 8:39 AM, Jeff Zhang <zjf...@gmail.com> wrote: > > > It seems you have connected to the right hadoop when you start pig > > grunt. But connect to the wrong hadoop when you run pig script. > > Try to search whether there's other configuration files that mess up > > with your default configuration. And what is machine "192.168.230.130" > > ? > > > > > > On Tue, Sep 28, 2010 at 9:23 AM, Ngô Văn Vĩ <ngovi.se....@gmail.com> > > wrote: > > > have you help me? > > > i have configuration > > > *- bin/pig* > > > export JAVA_HOME=/home/ngovi/jdk1.6.0_21 > > > export PIG_INSTALL=/home/ngovi/pig-0.7.0 > > > export PATH=$PATH:$PIG_INSTALL/bin > > > export PIG_HADOOP_VERSION=0.20.2 > > > export PIG_CLASSPATH=/home/ngovi/hadoop-0.20.2/conf/ > > > .... > > > *- conf/pig.properties* > > > fs.default.name=hdfs://localhost:9000/ > > > mapred.job.tracker=localhost:9001 > > > # log4jconf log4j configuration file > > > i run pig that have error > > > > > > *- in hadoop-0.20.2/conf* > > > *core-site.xml* > > > <configuration> > > > <property> > > > <name>fs.default.name</name> > > > <value>hdfs://localhost:9000</value> > > > <description> > > > the name of the default file system > > > </description> > > > </property> > > > </configuration> > > > *hdfs-site.xml* > > > <configuration> > > > <property> > > > <name>dfs.replication</name> > > > <value>1</value> > > > <description>Default block replication </description> > > > </property> > > > </configuration> > > > > > > *mapred-site.xml* > > > > > > <configuration> > > > <property> > > > <name>mapred.job.tracker</name> > > > <value>localhost:9001</value> > > > <description> > > > the host and port that the mapreduce job tracker run at > > > </description> > > > </property> > > > </configuration> > > > > > > I run pig that have error?? > > > *ng...@master:~/pig-0.7.0$ bin/pig -x mapreduce > > > 10/09/27 18:16:29 INFO pig.Main: Logging error messages to: > > > /home/ngovi/pig-0.7.0/pig_1285636589590.log > > > 2010-09-27 18:16:30,029 [main] INFO > > > org.apache.pig.backend.hadoop.executionengine.HExecutionEngine - > > Connecting > > > to hadoop file system at: hdfs://localhost:9000/ > > > 2010-09-27 18:16:30,347 [main] INFO > > > org.apache.pig.backend.hadoop.executionengine.HExecutionEngine - > > Connecting > > > to map-reduce job tracker at: localhost:9001 > > > grunt> * > > > > > > > > > thanks all > > > > > > On Mon, Sep 27, 2010 at 1:14 PM, Alan Gates <ga...@yahoo-inc.com> > wrote: > > > > > >> Pig is failing to connect to your namenode. Is the address Pig is > > trying > > >> to use (hdfs://master:54310/) correct? Can you connect using that > > string > > >> from the same machine using bin/hadoop? > > >> > > >> Alan. > > >> > > >> > > >> On Sep 27, 2010, at 8:45 AM, Ngô Văn Vĩ wrote: > > >> > > >> I run Pig at Hadoop Mode > > >>> (Pig-0.7.0 and hadoop-0.20.2) > > >>> have error? > > >>> ng...@master:~/pig-0.7.0$ bin/pig > > >>> 10/09/27 08:39:40 INFO pig.Main: Logging error messages to: > > >>> /home/ngovi/pig-0.7.0/pig_1285601980268.log > > >>> 2010-09-27 08:39:40,538 [main] INFO > > >>> org.apache.pig.backend.hadoop.executionengine.HExecutionEngine - > > >>> Connecting > > >>> to hadoop file system at: hdfs://master:54310/ > > >>> 2010-09-27 08:39:41,760 [main] INFO org.apache.hadoop.ipc.Client - > > >>> Retrying > > >>> connect to server: master/192.168.230.130:54310. Already tried 0 > > time(s). > > >>> 2010-09-27 08:39:42,762 [main] INFO org.apache.hadoop.ipc.Client - > > >>> Retrying > > >>> connect to server: master/192.168.230.130:54310. Already tried 1 > > time(s). > > >>> 2010-09-27 08:39:43,763 [main] INFO org.apache.hadoop.ipc.Client - > > >>> Retrying > > >>> connect to server: master/192.168.230.130:54310. Already tried 2 > > time(s). > > >>> 2010-09-27 08:39:44,765 [main] INFO org.apache.hadoop.ipc.Client - > > >>> Retrying > > >>> connect to server: master/192.168.230.130:54310. Already tried 3 > > time(s). > > >>> 2010-09-27 08:39:45,766 [main] INFO org.apache.hadoop.ipc.Client - > > >>> Retrying > > >>> connect to server: master/192.168.230.130:54310. Already tried 4 > > time(s). > > >>> 2010-09-27 08:39:46,767 [main] INFO org.apache.hadoop.ipc.Client - > > >>> Retrying > > >>> connect to server: master/192.168.230.130:54310. Already tried 5 > > time(s). > > >>> 2010-09-27 08:39:47,768 [main] INFO org.apache.hadoop.ipc.Client - > > >>> Retrying > > >>> connect to server: master/192.168.230.130:54310. Already tried 6 > > time(s). > > >>> 2010-09-27 08:39:48,769 [main] INFO org.apache.hadoop.ipc.Client - > > >>> Retrying > > >>> connect to server: master/192.168.230.130:54310. Already tried 7 > > time(s). > > >>> 2010-09-27 08:39:49,770 [main] INFO org.apache.hadoop.ipc.Client - > > >>> Retrying > > >>> connect to server: master/192.168.230.130:54310. Already tried 8 > > time(s). > > >>> 2010-09-27 08:39:50,771 [main] INFO org.apache.hadoop.ipc.Client - > > >>> Retrying > > >>> connect to server: master/192.168.230.130:54310. Already tried 9 > > time(s). > > >>> 2010-09-27 08:39:50,780 [main] ERROR org.apache.pig.Main - ERROR > 2999: > > >>> Unexpected internal error. Failed to create DataStorage > > >>> > > >>> Help me?? > > >>> Thanks > > >>> -- > > >>> Ngô Văn Vĩ > > >>> Công Nghệ Phần Mềm > > >>> Phone: 01695893851 > > >>> > > >> > > >> > > > > > > > > > -- > > > Ngô Văn Vĩ > > > Công Nghệ Phần Mềm > > > Phone: 01695893851 > > > > > > > > > > > -- > > Best Regards > > > > Jeff Zhang > > > > > > -- > Ngô Văn Vĩ > Công Nghệ Phần Mềm > Phone: 01695893851 >