It's weird. I doubt maybe there's other configuration file on your class path which override your real conf files. Could you download a new pig release and follow the instructions on http://hadoop.apache.org/pig/docs/r0.7.0/setup.html on a new environment ?
On Thu, Aug 26, 2010 at 7:49 PM, rahul <rmalv...@apple.com> wrote: > Hi , > > I tried the grunt shell as well but that also does not connects to hadoop. It > throws a warning and runs the job in standalone mode. So it tried it using > the pig.jar. > > Do you have any further suggestion on that ? > > Rahul > > On Aug 26, 2010, at 7:23 PM, Jeff Zhang wrote: > >> Connect to 9001 is right, this is jobtracker's ipc port while 50030 >> is its http server port. >> And have you ever try to run the grunt shell ? >> >> On Thu, Aug 26, 2010 at 7:12 PM, rahul <rmalv...@apple.com> wrote: >>> Hi Jeff, >>> >>> I can connect to the jobtracker web UI using the following URL : >>> http://localhost:50030/jobtracker.jsp >>> >>> And also I can see jobs which I ran directly using the streaming api on >>> hadoop. >>> >>> I also see it tries to connect to localhost/127.0.0.1:9001 which I have >>> specified in the hadoop conf file >>> and I have also tried changing this location to localhost:50030 but still >>> the error remains the same. >>> >>> Can you suggest something further ? >>> >>> Thanks, >>> Rahul >>> >>> On Aug 26, 2010, at 7:07 PM, Jeff Zhang wrote: >>> >>>> Can you look at the jobtracker log or access jobtracker web ui ? >>>> It seems you can not connect to jobtracker according your log >>>> >>>> "Caused by: java.io.IOException: Call to localhost/127.0.0.1:9001 >>>> failed on local exception: java.io.EOFException" >>>> >>>> >>>> >>>> On Fri, Aug 27, 2010 at 10:00 AM, rahul <rmalv...@apple.com> wrote: >>>>> Yes they are running. >>>>> >>>>> On Aug 26, 2010, at 6:59 PM, Jeff Zhang wrote: >>>>> >>>>>> Execute command jps in shell to see whether namenode and jobtracker is >>>>>> running correctly. >>>>>> >>>>>> >>>>>> >>>>>> On Fri, Aug 27, 2010 at 9:49 AM, rahul <rmalv...@apple.com> wrote: >>>>>>> Hi Jeff, >>>>>>> >>>>>>> I transferred the hadoop conf files to the pig/conf location but still >>>>>>> i get the same error. >>>>>>> >>>>>>> Does the issue is with the configuration files or with the hdfs files >>>>>>> system ? >>>>>>> >>>>>>> Can test the connection to hdfs(localhost/127.0.0.1:9001) in some way ? >>>>>>> >>>>>>> Steps I did : >>>>>>> >>>>>>> 1. I have formatted initially my local file system using the ./hadoop >>>>>>> namenode -format command. I believe this mounts the local file system >>>>>>> to HDFS. >>>>>>> 2. Then I configured the hadoop conf files and started ./start-all >>>>>>> script. >>>>>>> 3. Started Pig with a custom pig script which should read hdfs as I >>>>>>> passed the HADOOP_CONF_DIR as parameter. >>>>>>> The command was java -cp $PIGDIR/pig.jar:$HADOOP_CONF_DIR >>>>>>> org.apache.pig.Main script1-hadoop.pig >>>>>>> >>>>>>> Please let me know if these step miss something ? >>>>>>> >>>>>>> Thanks, >>>>>>> Rahul >>>>>>> >>>>>>> >>>>>>> On Aug 26, 2010, at 6:33 PM, Jeff Zhang wrote: >>>>>>> >>>>>>>> Try to put the hadoop xml configuration file to pig/conf folder >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Thu, Aug 26, 2010 at 6:22 PM, rahul <rmalv...@apple.com> wrote: >>>>>>>>> Hi Jeff, >>>>>>>>> >>>>>>>>> I have set the hadoop conf in class path by setting $HADOOP_CONF_DIR >>>>>>>>> variable. >>>>>>>>> >>>>>>>>> But I have both Pig and hadoop running at the same machine, so >>>>>>>>> localhost should not make a difference. >>>>>>>>> >>>>>>>>> So I have used all the default config setting for the core-site.xml, >>>>>>>>> hdfs-site.xml, mapred-site.xml, as per the hadoop tutorial. >>>>>>>>> >>>>>>>>> Please let me know if my understanding is correct ? >>>>>>>>> >>>>>>>>> I am attaching the conf files as well : >>>>>>>>> hdfs-site.xml: >>>>>>>>> >>>>>>>>> <?xml version="1.0"?> >>>>>>>>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >>>>>>>>> >>>>>>>>> <!-- Put site-specific property overrides in this file. --> >>>>>>>>> >>>>>>>>> <configuration> >>>>>>>>> <property> >>>>>>>>> <name>fs.default.name</name> >>>>>>>>> <value>hdfs://localhost:9000</value> >>>>>>>>> <description>The name of the default file system. A URI whose >>>>>>>>> scheme and authority determine the FileSystem implementation. The >>>>>>>>> uri's scheme determines the config property (fs.SCHEME.impl) naming >>>>>>>>> the FileSystem implementation class. The uri's authority is used to >>>>>>>>> determine the host, port, etc. for a filesystem.</description> >>>>>>>>> </property> >>>>>>>>> >>>>>>>>> <property> >>>>>>>>> <name>dfs.replication</name> >>>>>>>>> <value>1</value> >>>>>>>>> <description>Default block replication. >>>>>>>>> The actual number of replications can be specified when the file is >>>>>>>>> created. >>>>>>>>> The default is used if replication is not specified in create time. >>>>>>>>> </description> >>>>>>>>> </property> >>>>>>>>> >>>>>>>>> </configuration> >>>>>>>>> >>>>>>>>> core-site.xml >>>>>>>>> <?xml version="1.0"?> >>>>>>>>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >>>>>>>>> >>>>>>>>> <!-- Put site-specific property overrides in this file. --> >>>>>>>>> >>>>>>>>> <configuration> >>>>>>>>> <property> >>>>>>>>> <name>hadoop.tmp.dir</name> >>>>>>>>> <value>/Users/rahulmalviya/Documents/Hadoop/hadoop-0.21.0/hadoop-${user.name}</value> >>>>>>>>> <description>A base for other temporary directories.</description> >>>>>>>>> </property> >>>>>>>>> </configuration> >>>>>>>>> >>>>>>>>> mapred-site.xml >>>>>>>>> <?xml version="1.0"?> >>>>>>>>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >>>>>>>>> >>>>>>>>> <!-- Put site-specific property overrides in this file. --> >>>>>>>>> >>>>>>>>> <configuration> >>>>>>>>> <property> >>>>>>>>> <name>mapred.job.tracker</name> >>>>>>>>> <value>localhost:9001</value> >>>>>>>>> <description>The host and port that the MapReduce job tracker runs >>>>>>>>> at. If "local", then jobs are run in-process as a single map >>>>>>>>> and reduce task. >>>>>>>>> </description> >>>>>>>>> </property> >>>>>>>>> >>>>>>>>> <property> >>>>>>>>> <name>mapred.tasktracker.tasks.maximum</name> >>>>>>>>> <value>8</value> >>>>>>>>> <description>The maximum number of tasks that will be run >>>>>>>>> simultaneously by a >>>>>>>>> a task tracker >>>>>>>>> </description> >>>>>>>>> </property> >>>>>>>>> </configuration> >>>>>>>>> >>>>>>>>> Please let me know if there is a issue in my configurations ? Any >>>>>>>>> input is valuable for me. >>>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> Rahul >>>>>>>>> >>>>>>>>> On Aug 26, 2010, at 6:10 PM, Jeff Zhang wrote: >>>>>>>>> >>>>>>>>>> Do you put the hadoop conf on classpath ? It seems you are still >>>>>>>>>> using >>>>>>>>>> local file system but conncect Hadoop's JobTracker. >>>>>>>>>> Make sure you set the correct configuration in core-site.xml >>>>>>>>>> hdfs-site.xml, mapred-site.xml, and put them on classpath. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Thu, Aug 26, 2010 at 5:32 PM, rahul <rmalv...@apple.com> wrote: >>>>>>>>>>> Hi , >>>>>>>>>>> >>>>>>>>>>> I am trying to integrate Pig with Hadoop for processing of jobs. >>>>>>>>>>> >>>>>>>>>>> I am able to run Pig in local mode and Hadoop with streaming api >>>>>>>>>>> perfectly. >>>>>>>>>>> >>>>>>>>>>> But when I try to run Pig with Hadoop I get follwong Error: >>>>>>>>>>> >>>>>>>>>>> Pig Stack Trace >>>>>>>>>>> --------------- >>>>>>>>>>> ERROR 2116: Unexpected error. Could not validate the output >>>>>>>>>>> specification for: >>>>>>>>>>> file:///Users/rahulmalviya/Documents/Pig/dev/main_merged_hdp_out >>>>>>>>>>> >>>>>>>>>>> org.apache.pig.impl.plan.PlanValidationException: ERROR 0: An >>>>>>>>>>> unexpected exception caused the validation to stop >>>>>>>>>>> at >>>>>>>>>>> org.apache.pig.impl.plan.PlanValidator.validate(PlanValidator.java:56) >>>>>>>>>>> at >>>>>>>>>>> org.apache.pig.impl.logicalLayer.validators.InputOutputFileValidator.validate(InputOutputFileValidator.java:49) >>>>>>>>>>> at >>>>>>>>>>> org.apache.pig.impl.logicalLayer.validators.InputOutputFileValidator.validate(InputOutputFileValidator.java:37) >>>>>>>>>>> at >>>>>>>>>>> org.apache.pig.impl.logicalLayer.validators.LogicalPlanValidationExecutor.validate(LogicalPlanValidationExecutor.java:89) >>>>>>>>>>> at org.apache.pig.PigServer.validate(PigServer.java:930) >>>>>>>>>>> at org.apache.pig.PigServer.compileLp(PigServer.java:910) >>>>>>>>>>> at org.apache.pig.PigServer.compileLp(PigServer.java:871) >>>>>>>>>>> at org.apache.pig.PigServer.compileLp(PigServer.java:852) >>>>>>>>>>> at org.apache.pig.PigServer.execute(PigServer.java:816) >>>>>>>>>>> at org.apache.pig.PigServer.access$100(PigServer.java:105) >>>>>>>>>>> at >>>>>>>>>>> org.apache.pig.PigServer$Graph.execute(PigServer.java:1080) >>>>>>>>>>> at org.apache.pig.PigServer.executeBatch(PigServer.java:288) >>>>>>>>>>> at >>>>>>>>>>> org.apache.pig.tools.grunt.GruntParser.executeBatch(GruntParser.java:109) >>>>>>>>>>> at >>>>>>>>>>> org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:166) >>>>>>>>>>> at >>>>>>>>>>> org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:138) >>>>>>>>>>> at org.apache.pig.tools.grunt.Grunt.exec(Grunt.java:89) >>>>>>>>>>> at org.apache.pig.Main.main(Main.java:391) >>>>>>>>>>> Caused by: org.apache.pig.impl.plan.PlanValidationException: ERROR >>>>>>>>>>> 2116: Unexpected error. Could not validate the output specification >>>>>>>>>>> for: >>>>>>>>>>> file:///Users/rahulmalviya/Documents/Pig/dev/main_merged_hdp_out >>>>>>>>>>> at >>>>>>>>>>> org.apache.pig.impl.logicalLayer.validators.InputOutputFileVisitor.visit(InputOutputFileVisitor.java:93) >>>>>>>>>>> at >>>>>>>>>>> org.apache.pig.impl.logicalLayer.LOStore.visit(LOStore.java:140) >>>>>>>>>>> at >>>>>>>>>>> org.apache.pig.impl.logicalLayer.LOStore.visit(LOStore.java:37) >>>>>>>>>>> at >>>>>>>>>>> org.apache.pig.impl.plan.DepthFirstWalker.depthFirst(DepthFirstWalker.java:67) >>>>>>>>>>> at >>>>>>>>>>> org.apache.pig.impl.plan.DepthFirstWalker.depthFirst(DepthFirstWalker.java:69) >>>>>>>>>>> at >>>>>>>>>>> org.apache.pig.impl.plan.DepthFirstWalker.depthFirst(DepthFirstWalker.java:69) >>>>>>>>>>> at >>>>>>>>>>> org.apache.pig.impl.plan.DepthFirstWalker.walk(DepthFirstWalker.java:50) >>>>>>>>>>> at >>>>>>>>>>> org.apache.pig.impl.plan.PlanVisitor.visit(PlanVisitor.java:51) >>>>>>>>>>> at >>>>>>>>>>> org.apache.pig.impl.plan.PlanValidator.validate(PlanValidator.java:50) >>>>>>>>>>> ... 16 more >>>>>>>>>>> Caused by: java.io.IOException: Call to localhost/127.0.0.1:9001 >>>>>>>>>>> failed on local exception: java.io.EOFException >>>>>>>>>>> at >>>>>>>>>>> org.apache.hadoop.ipc.Client.wrapException(Client.java:775) >>>>>>>>>>> at org.apache.hadoop.ipc.Client.call(Client.java:743) >>>>>>>>>>> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220) >>>>>>>>>>> at >>>>>>>>>>> org.apache.hadoop.mapred.$Proxy0.getProtocolVersion(Unknown Source) >>>>>>>>>>> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359) >>>>>>>>>>> at >>>>>>>>>>> org.apache.hadoop.mapred.JobClient.createRPCProxy(JobClient.java:429) >>>>>>>>>>> at >>>>>>>>>>> org.apache.hadoop.mapred.JobClient.init(JobClient.java:423) >>>>>>>>>>> at >>>>>>>>>>> org.apache.hadoop.mapred.JobClient.<init>(JobClient.java:410) >>>>>>>>>>> at org.apache.hadoop.mapreduce.Job.<init>(Job.java:50) >>>>>>>>>>> at >>>>>>>>>>> org.apache.pig.impl.logicalLayer.validators.InputOutputFileVisitor.visit(InputOutputFileVisitor.java:89) >>>>>>>>>>> ... 24 more >>>>>>>>>>> Caused by: java.io.EOFException >>>>>>>>>>> at java.io.DataInputStream.readInt(DataInputStream.java:375) >>>>>>>>>>> at >>>>>>>>>>> org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:501) >>>>>>>>>>> at >>>>>>>>>>> org.apache.hadoop.ipc.Client$Connection.run(Client.java:446) >>>>>>>>>>> ================================================================================ >>>>>>>>>>> >>>>>>>>>>> Did anyone got the same error. I think it related to connection >>>>>>>>>>> between pig and hadoop. >>>>>>>>>>> >>>>>>>>>>> Can someone tell me how to connect Pig and hadoop. >>>>>>>>>>> >>>>>>>>>>> Thanks. >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Best Regards >>>>>>>>>> >>>>>>>>>> Jeff Zhang >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Best Regards >>>>>>>> >>>>>>>> Jeff Zhang >>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Best Regards >>>>>> >>>>>> Jeff Zhang >>>>> >>>>> >>>> >>>> >>>> >>>> -- >>>> Best Regards >>>> >>>> Jeff Zhang >>> >>> >> >> >> >> -- >> Best Regards >> >> Jeff Zhang > > -- Best Regards Jeff Zhang