Creating custom input split.l

2011-04-06 Thread ranjith k
Hello. I need to create a custom input split. I need to split my input in to 50 line for one input split. How can i do it. And also there is an another problem for me. I have a file. But it is not in the form of text. It contain structure. I need to give one structure in to my map function as valu

Re: hadoop cluster install path

2011-04-06 Thread zhengjun chen
I did as you said. I built a small hadoop cluster. It consists of three nodes. One is master, two slaves. The master node is name node and jobtracker. The other two nodes are data nodes and tasktracker. Different nodes installs hadoop on different paths. I run map/reduce program. The program can w

Re: Need 0.20.2 new API documentation/examples, where are they?

2011-04-06 Thread Aaron Kimball
Simplest answer: Job A uses o.a.h.mapreduce.lib.output.SequenceFileOutputFormat It writes values to that (using context.write()) of classes KT, VT Use o.a.h.mapreduce.lib.output.FileOutputFormat.setOutputPath(job, new Path("job-a-out")); to configure the job to write to some location. Then run jo

Distributed Cache File Not Found Exception

2011-04-06 Thread Shuja Rehman
Hi, I m getting FileNotFound Exception while using distribute cache. here is the details. Configuration config = new Configuration(); config.clear(); config.set("hbase.zookeeper.quorum", HBASE_ZOOKEEPER_QUORUM); config.set("hbase.zookeeper.property.clientPort", HBASE_ZOOKEEPER_PROPERTY_CLIENTPORT)

Terasort error when running Hadoop over PVFS2

2011-04-06 Thread Wantao
Hi guys, I am trying to run Hadoop0.21.0 with PVFS2. Following the email thread http://www.mail-archive.com/core-user@hadoop.apache.org/msg04434.html, I managed to make my Hadoop cluster up and run wordcount example successfully. Then I plan to perform some test for Terasort example. The data g

error in running the WordCount

2011-04-06 Thread لسٹ शिराज़
Hi, I am a newbie to the mapreduce (in fact hadoop as a whole) framework. I am trying to run a simple WordCount client class programatically inside the eclipse, hence for that, I have provided the configuration files for hdfs, mapreduce and core using the JobConf.addResource method respectively. He